N - Channel Allocation

N - Channel Allocation

 

#include<stdio.h>
#include<string.h>
int main()
{
	int n, map[27], book[27][27], i, j;
	char a[27];
	while(~scanf("%d", &n), n)
	{
		memset(book, 0, sizeof(book));
		memset(map, 0, sizeof(map));
		for(i=1;i<=n;i++)
		{
			scanf("%s", a);
			for(j=2;a[j]!='\0';j++)
				book[i][a[j]-'A'+1]=1;
		}
		int c[27]={0};
		for(i=1;i<=n;i++)
		{
			memset(c, 0, sizeof(c));
			for(j=1;j<=n;j++)
				if(book[i][j])
					c[map[j]]=1;
			for(j=1;j<=n;j++)
				if(c[j]==0)
				{
					map[i]=j;
					break;
				}
		}		
		int temp=0;
		for(i=1;i<=n;i++)
			if(map[i]>temp)
				temp=map[i];
		if(temp>1)
			printf("%d channels needed.\n", temp);
		else
			printf("%d channel needed.\n", temp);
	}
	return 0;
}
  
  

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论
Recall that to solve (P2) in the tth time frame, we observe ξt 􏰗 {hti, Qi(t), Yi(t)}Ni=1, consisting of the channel gains {hti}Ni=1 and the system queue states {Qi(t),Yi(t)}Ni=1, and accordingly decide the control action {xt, yt}, including the binary offloading decision xt and the continuous resource allocation yt 􏰗 􏰄τit, fit, eti,O, rit,O􏰅Ni=1. A close observation shows that although (P2) is a non-convex optimization problem, the resource allocation problem to optimize yt is in fact an “easy” convex problem if xt is fixed. In Section IV.B, we will propose a customized algorithm to efficiently obtain the optimal yt given xt in (P2). Here, we denote G􏰀xt,ξt􏰁 as the optimal value of (P2) by optimizing yt given the offloading decision xt and parameter ξt. Therefore, solving (P2) is equivalent to finding the optimal offloading decision (xt)∗, where (P3) : 􏰀xt􏰁∗ = arg maximize G 􏰀xt, ξt􏰁 . (20) xt ∈{0,1}N In general, obtaining (xt)∗ requires enumerating 2N offloading decisions, which leads to significantly high computational complexity even when N is moderate (e.g., N = 10). Other search based methods, such as branch-and-bound and block coordinate descent [29], are also time-consuming when N is large. In practice, neither method is applicable to online decision- making under fast-varying channel condition. Leveraging the DRL technique, we propose a LyDROO algorithm to construct a policy π that maps from the input ξt to the optimal action (xt)∗, i.e., π : ξt 􏰕→ (xt)∗, with very low complexity, e.g., tens of milliseconds computation time (i.e., the time duration from observing ξt to producing a control action {xt, yt}) when N = 10深度强化学习的动作是什么
最新发布
06-03

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

hhjian6666

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值