题目大意:
n个点,合并成1个点,每次合并不超过k个点,合并的花费是合并点的价值和,合并之后生成的点的价值也是合并点的价值和。
问在满足花费 <= T的条件下,最小的k
二分很好想到。
考虑什么情况下能每次k个合并恰好合并完。
即为n+x*(k-1) == 1
n-1 == x*(k-1)
(n-1)%(k-1) == 0的情况
其余情况会出现(n-1)%(k-1) = r,那么就要有一次合并r个节点,为了花费尽量小,在总的合并次数固定的情况下,尽量让小的多合并几次,大的少合并几次,那么就把多余的次数在第一次合并掉,就避免了累赘在最后。
较好的处理方法是补0,补齐到(n-1)%(k-1) == 0
然后合并不能单纯用优先队列,会TLE。价值 <= 1000的用数组模拟。
代码如下:
#include <iostream>
#include <cmath>
#include <vector>
#include <cstdlib>
#include <cstdio>
#include <climits>
#include <ctime>
#include <cstring>
#include <queue>
#include <stack>
#include <list>
#include <algorithm>
#include <map>
#include <set>
#define LL long long
#define Pr pair<int,int>
#define fread(ch) freopen(ch,"r",stdin)
#define fwrite(ch) freopen(ch,"w",stdout)
using namespace std;
const int INF = 0x3f3f3f3f;
const int msz = 10000;
const int mod = 1e9+7;
const double eps = 1e-8;
int n,mx;
int num[112345];
int cnt[1010];
priority_queue <int,vector<int>,greater<int> > q;
LL cal(int k)
{
memset(cnt,0,sizeof(cnt));
while(!q.empty()) q.pop();
int tmp = (n-1+k-2)/(k-1);
tmp = tmp*(k-1)+1;
//printf("*%d %d\n",k,tmp);
int low = 0;
for(int i = 0; i < n; ++i)
{
if(num[i] <= 1000) cnt[num[i]]++;
else q.push(num[i]);
}
for(int i = n; i < tmp; ++i)
cnt[0]++;
LL ans = 0;
while(tmp != 1)
{
int us = 0;
int x = 0;
for(int &i = low; i <= 1000; ++i)
{
if(us+cnt[i] >= k)
{
cnt[i] -= k-us;
//printf("!%d %d\n",k-us,i);
x += i*(k-us);
us = k;
break;
}
x += cnt[i]*i;
//printf("!%d %d\n",cnt[i],i);
us += cnt[i];
cnt[i] = 0;
}
while(us < k)
{
x += q.top();
q.pop();
us++;
}
tmp -= k-1;
ans += x;
//printf("%d %lld\n",x,ans);
if(x <= 1000) cnt[x]++;
else q.push(x);
}
//printf("%d %lld %d\n",k,ans,mx);
return ans;
}
int solve()
{
int k;
int l,r;
l = 2,r = n;
while(l <= r)
{
int mid = (l+r)>>1;
if(cal(mid) <= mx)
{
k = mid;
r = mid-1;
}
else l = mid+1;
}
return k;
}
int main()
{
//fread("");
//fwrite("");
int t;
scanf("%d",&t);
while(t--)
{
scanf("%d%d",&n,&mx);
for(int i = 0; i < n; ++i) scanf("%d",&num[i]);
printf("%d\n",solve());
}
return 0;
}