不知道为啥,最近都是二维前缀和。昨天说的那两个都是一维的前缀和,今天需要解决二维的前缀和,暴力解决是O(n ^ 4)。我们现在想通过利用数据结构将时间复杂度从O(n ^ 4)降到O(n ^ 3 *logn)。昨天那个一维的也是这样的,通过枚举右端点,同时记录之前的值,看作查表。参考宫水三叶姐姐的题解,赞 !
363 矩形区域不超过 K 的最大数值和
class Solution {
public:
int maxSumSubmatrix(vector<vector<int>>& matrix, int k) {
int m = matrix.size();
int n = matrix[0].size();
int ans = INT_MIN;
int s[m + 1][n + 1];
memset(s, 0, sizeof s);
for(int i = 1; i <= m; i++)
for(int j = 1; j <= n; j++)
s[i][j] = s[i - 1][j] + s[i][j - 1] - s[i - 1][j - 1] + matrix[i - 1][j - 1];
for(int x1 = 1; x1 <= m; x1++)
{
for(int x2 = x1; x2 <= m; x2++)
{
set<int> st;
st.insert(0);
for(int y2 = 1; y2 <= n; y2++)
{
int A = s[x2][y2] - s[x1 - 1][y2];
auto left = st.lower_bound(A - k);//不小于A-k的第一个元素
if(left != st.end())
{
int t = A - *left;
ans = max(ans, t);
}
st.insert(A);
}
}
}
return ans;
}
};
1074 元素和为目标值的子矩阵数量
同样的,不过这次需要统计数量,要把数量也保存起来
class Solution {
public:
int numSubmatrixSumTarget(vector<vector<int>>& matrix, int target) {
int m = matrix.size();
int n = matrix[0].size();
int ans = 0;
int s[m + 1][n + 1];
memset(s, 0, sizeof s);
for(int i = 1; i <= m; i++)
for(int j = 1; j <= n; j++)
s[i][j] = s[i - 1][j] + s[i][j - 1] - s[i - 1][j - 1] + matrix[i - 1][j - 1];
for(int x1 = 1; x1 <= m; x1++)
{
for(int x2 = x1; x2 <= m; x2++)
{
unordered_map<int, int> st;
st[0] = 1;
for(int y2 = 1; y2 <= n; y2++)
{
int A = s[x2][y2] - s[x1 - 1][y2];
if(st.find(A - target) != st.end())
ans += st[A - target];
if(st.find(A) != st.end())
st[A]++;
else
st[A] = 1;
}
}
}
return ans;
}
};