项目简介:
项目功能说明:
- 该网站包括file页面、transfer页面和settings页面。其中,file页面用于显示指定目录下的所有文件,可根据文件类型、名称或者大小进行排序筛选;transfer页面主要的使用场景是实现提交HTTP地址后下载到服务器上,这个服务器可以作为NAS或者云盘使用;settings页面用于设置下载路径、限速大小、最大并行任务数。
- transfer页面能够查看每个任务的下载进度、下载速度和下载剩余时间,用户可自行控制并发线程数。
技术栈: Gin + Gorm + Sqlite + SSE
项目地址(欢迎star):
JAVA版本:https://gitee.com/Liguangyu1017/HTTP-download-server
GO版本: https://github.com/Lemon001017/HTTP-download-server
JAVA版本的实现细节见: http://t.csdnimg.cn/1T1Km
效果展示:

API 文档:
查看方式:
项目运行后浏览器输入:http://localhost:8000/api/docs/

分片下载实现步骤:
- 用户提交下载 url
- 获取文件大小、保存路径、最大速度等
- 计算分片大小、分片数量,初始化所有分片
- 创建 goroutine 池,循环将所有分片丢入池中 (调用ants第三方库)
- 动态计算下载速度、进度、剩余时间,并通过 sse 发送到客户端
- 等待下载结束
1. 获取下载配置
func (h *Handlers) getSettingsInfo() (string, float64, uint, error) {
settings, err := models.GetSettings(h.db, 1)
if err != nil {
return "", 0, 0, err
}
outputDir := settings.DownloadPath
if outputDir == "" {
outputDir, err = os.Getwd()
if err != nil {
return "", 0, 0, err
}
}
maxDownloadSpeed := settings.MaxDownloadSpeed
if maxDownloadSpeed == 0 {
maxDownloadSpeed = 1e9
}
maxTasks := settings.MaxTasks
if maxTasks == 0 {
maxTasks = 4
}
return outputDir, maxDownloadSpeed, maxTasks, nil
}
2. 计算分片大小、数量
分片策略:
本项目采用固定分片大小,即根据文件大小来决定分片数量。
根据文件大小,文件大小小于等于10MB的按照32KB,10MB~100MB按照1MB,超过100MB按照10MB进行分片。
func (h *Handlers) getChunkInfo(fileSize int64) (int64, int64) {
var chunkSize int64
switch {
case fileSize <= 10*1024*1024:
chunkSize = models.MinChunkSize
case fileSize <= 100*1024*1024:
chunkSize = models.MidChunkSize
default:
chunkSize = models.MaxChunkSize
}
numChunks := (fileSize + chunkSize - 1) / chunkSize
return chunkSize, numChunks
}
3. 获取文件信息
func (h *Handlers) getFileInfo(url string, outputDir string) (int64, string, string, error) {
req, err := http.NewRequest("GET", url, nil)
if err != nil {
return 0, "", "", err
}
req.Header.Set("User-Agent", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.159 Safari/537.36")
req.Header.Set("Accept", "*/*")
client := &http.Client{}
resp, err := client.Do(req)
if err != nil {
return 0, "", "", err
}
defer resp.Body.Close()
if resp.StatusCode != http.StatusOK {
return 0, "", "", err
}
fileSize := resp.ContentLength
fileName := extractFileName(resp, url)
outputPath := filepath.Join(outputDir, fileName)
return fileSize, outputPath, fileName, nil
}
func extractFileName(resp *http.Response, downloadURL string) string {
if contentDisposition := resp.Header.Get("Content-Disposition"); contentDisposition != "" {
_, params, err := mime.ParseMediaType(contentDisposition)
if err == nil && params["filename"] != "" {
return params["filename"]
}
}
parsedURL, err := url.Parse(downloadURL)
if err != nil {
parsedURL.Path = "/unknown"
}
re := regexp.MustCompile(`[^\/]+\.[a-zA-Z0-9]+$`)
fileName := re.FindString(parsedURL.Path)
if fileName == "" {
fileName = "unknown_file"
}
return fileName
}
4. 初始化一个下载任务
func (h *Handlers) initOneTask(url, key string) (*models.Task, error) {
outputDir, _, _, err := h.getSettingsInfo()
if err != nil {
return nil, err
}
fileSize, outputPath, fileName, err := h.getFileInfo(url, outputDir)
if err != nil {
return nil, err
}
chunkSize, numChunks := h.getChunkInfo(fileSize)
task := models.Task{
ID: key,
Name: fileName,
Url: url,
Size: fileSize,
SavePath: outputPath,
FileType: filepath.Ext(fileName),
Threads: models.DefaultThreads,
Status: models.TaskStatusPending,
ChunkNum: numChunks,
ChunkSize: chunkSize,
Chunk: make([]models.Chunk, numChunks),
}
err = models.AddTask(h.db, &task)
if err != nil {
return nil, err
}
return &task, nil
}
5. 提交下载任务,异步执行下载,动态计算下载数据并发送给客户端
注意:每个分片下载完成之后,存储到分片的起始位置中,这个过程由于文件句柄只有一个,需要加锁(mutex)后进行seek和write,保证存储位置的准确性。
// submit task
func (h *Handlers) handleSubmit(c *gin.Context) {
var request DownloadRequest
err := c.ShouldBindJSON(&request)
if err != nil {
carrot.AbortWithJSONError(c, http.StatusBadRequest, err)
return
}
eventSource := h.createEventSource()
task, err := h.initOneTask(request.URL, eventSource.key)
if err != nil {
carrot.AbortWithJSONError(c, http.StatusInternalServerError, err)
return
}
// Open a goroutine to handle the download separately
go func() {
h.processDownload(eventSource, task, 0)
}()
c.JSON(http.StatusOK, EventSourceResult{Key: eventSource.key})
}
func (h *Handlers) processDownload(es *EventSource, task *models.Task, lastTotalDownloaded int64) {
startTime := time.Now()
carrot.Info("id:", task.ID, "fileSize:", task.Size, "savePath:", task.SavePath, "chunkSize:", task.ChunkSize, "numChunks:", task.ChunkNum)
outputFile, err := os.OpenFile(task.SavePath, os.O_RDWR|os.O_CREATE, 0644)
if err != nil {
carrot.Error("open or create file error", "key:", es.key, "id:", task.ID, "url:", task.Url, "err:", err)
return
}
_, maxDownloadSpeed, _, err := h.getSettingsInfo()
if err != nil {
carrot.Error("get settings error", "key:", es.key, "id:", task.ID, "url:", task.Url, "err:", err)
return
}
// Init chunk info
if task.TotalDownloaded == 0 {
for i := 0; i < int(task.ChunkNum); i++ {
start := int64(i) * task.ChunkSize
end := math.Min(float64(start+task.ChunkSize), float64(task.Size)) - 1
task.Chunk[i] = models.Chunk{
TaskID: task.ID,
Index: i,
Start: int(start),
End: int(end),
}
models.AddChunk(h.db, &task.Chunk[i])
}
}
task.Status = models.TaskStatusDownloading
models.UpdateTask(h.db, task)
// Create a pool of goroutines
pool, _ := ants.NewPoolWithFunc(int(task.Threads), func(i interface{}) {
err := h.downloadChunk(es, &task.Chunk[i.(int)], task, outputFile, startTime, maxDownloadSpeed, lastTotalDownloaded)
if err != nil {
if !errors.Is(err, context.Canceled) && !errors.Is(err, os.ErrClosed) {
carrot.Error("download chunk failed", "key:", es.key, "url:", task.Url, "err:", err)
task.Status = models.TaskStatusFailed
models.UpdateTask(h.db, task)
es.Emit(DownloadProgress{ID: task.ID, Name: task.Name, Status: models.TaskStatusFailed})
}
// Clean all resources
outputFile.Close()
h.cleanEventSource(task.ID)
return
}
})
defer pool.Release()
for i := 0; i < int(task.ChunkNum); i++ {
// Skip completed chunks
if !task.Chunk[i].Done {
_ = pool.Invoke(i)
}
}
}
func (h *Handlers) downloadChunk(es *EventSource, chunk *models.Chunk, task *models.Task,
outputFile *os.File, startTime time.Time, maxDownloadSpeed float64, lastTotalDownloaded int64) error {
req, err := http.NewRequest(http.MethodGet, task.Url, nil)
if err != nil {
carrot.Error("Failed to create HTTP request", "key:", es.key, "url:", task.Url)
return err
}
req.Header.Set("User-Agent", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/127.0.0.0 Safari/537.36")
req.Header.Set("Accept", "*/*")
req.Header.Set("Range", fmt.Sprintf("bytes=%v-%v", chunk.Start, chunk.End))
req = req.WithContext(es.ctx)
resp, err := h.client.Do(req)
if err != nil {
return err
}
defer resp.Body.Close()
if resp.StatusCode != http.StatusPartialContent {
carrot.Error("key:", es.key, "url:", task.Url, "status:", resp.StatusCode)
return err
}
buf := make([]byte, 2048)
h.mu.Lock()
_, err = outputFile.Seek(int64(chunk.Start), 0)
if err != nil {
h.mu.Unlock()
return err
}
// Create a rate limiter
maxDownloadSpeedInBytes := maxDownloadSpeed * 1000 * 1000
limiter := rate.NewLimiter(rate.Limit(maxDownloadSpeedInBytes), int(maxDownloadSpeedInBytes*0.1))
lastMessageTime := time.Now()
for {
n, err := resp.Body.Read(buf)
if err != nil && err != io.EOF {
h.mu.Unlock()
return err
}
if n == 0 {
break
}
err = limiter.WaitN(es.ctx, n)
if err != nil {
h.mu.Unlock()
return err
}
_, err = outputFile.Write(buf[:n])
if err != nil {
h.mu.Unlock()
return err
}
task.TotalDownloaded += int64(n)
if time.Since(lastMessageTime) >= models.MessageInterval {
speed, progress, remainingTime := h.calculateDownloadData(task, startTime, lastTotalDownloaded)
es.Emit(DownloadProgress{
ID: task.ID,
Name: task.Name,
Progress: progress,
Speed: speed,
RemainingTime: remainingTime,
Status: task.Status,
})
models.UpdateTask(h.db, task)
lastMessageTime = time.Now()
}
}
chunk.Done = true
models.UpdateChunk(h.db, chunk)
// download complete
if task.TotalDownloaded == task.Size {
task.Status = models.TaskStatusDownloaded
models.UpdateTask(h.db, task)
es.Emit(DownloadProgress{
ID: task.ID,
Name: task.Name,
Status: task.Status,
})
carrot.Info("Download complete", "key:", es.key, "id:", task.ID, "url:", task.Url)
models.DeleteChunks(h.db, task.ID)
outputFile.Close()
close(es.eventChan)
} else if task.TotalDownloaded > task.Size {
return models.ErrExpectedFileSize
}
models.UpdateTask(h.db, task)
h.mu.Unlock()
return nil
}
6. 等待下载结束,判断是否下载成功,清理资源:
// download complete
if task.TotalDownloaded == task.Size {
task.Status = models.TaskStatusDownloaded
models.UpdateTask(h.db, task)
es.Emit(DownloadProgress{
ID: task.ID,
Name: task.Name,
Status: task.Status,
})
carrot.Info("Download complete", "key:", es.key, "id:", task.ID, "url:", task.Url)
models.DeleteChunks(h.db, task.ID)
outputFile.Close()
close(es.eventChan)
} else if task.TotalDownloaded > task.Size {
return models.ErrExpectedFileSize
}
暂停下载实现步骤:
- 获取所有 id
- 根据 id 获取所有任务
- 枚举所有任务,更新任务状态,然后调用 cleanEventSource 方法执行上下文取消
// Pause download
func (h *Handlers) handlePause(c *gin.Context) {
var ids []string
err := c.ShouldBindJSON(&ids)
if err != nil {
carrot.AbortWithJSONError(c, http.StatusBadRequest, err)
return
}
tasks, err := models.GetTaskByIds(h.db, ids)
if err != nil {
carrot.AbortWithJSONError(c, http.StatusInternalServerError, err)
return
}
for _, task := range tasks {
if task.Status == models.TaskStatusDownloading {
h.cleanEventSource(task.ID)
task.Status = models.TaskStatusCanceled
models.UpdateTask(h.db, &task)
} else {
carrot.AbortWithJSONError(c, http.StatusBadRequest, models.ErrStatusNotDownloading)
return
}
}
c.JSON(http.StatusOK, gin.H{"ids": ids})
}
// Clean resources
func (h *Handlers) cleanEventSource(key string) {
v, ok := h.eventSources.LoadAndDelete(key)
if !ok {
return
}
eventSource, ok := v.(*EventSource)
if !ok {
return
}
eventSource.cancel()
if eventSource.eventChan != nil {
close(eventSource.eventChan)
eventSource.eventChan = nil
}
}
恢复下载实现步骤:
关键: 每个下载任务都对应一个记分牌,用于记录 taskId、分片索引、开始位置、结束位置和是否下载完成,这些数据存入 chunks 表中;当这个分片下载完成就会标记为 true 并更新数据库,这样再次开启下载只需要提交未下载的分片到 pool 中即可。
- 获取所有 id
- 根据 id 获取所有任务
- 枚举所有任务,更新任务状态,根据任务 id 获取所有分片,异步调用 processDownload 方法执行下载
- 等待下载结束
// Resume download
func (h *Handlers) handleResume(c *gin.Context) {
var ids []string
err := c.ShouldBindJSON(&ids)
if err != nil {
carrot.AbortWithJSONError(c, http.StatusBadRequest, err)
return
}
tasks, err := models.GetTaskByIds(h.db, ids)
if err != nil {
carrot.AbortWithJSONError(c, http.StatusInternalServerError, err)
return
}
for _, task := range tasks {
if task.Status != models.TaskStatusCanceled {
carrot.AbortWithJSONError(c, http.StatusBadRequest, models.ErrStatusNotCanceled)
return
}
es := h.createEventSourceWithKey(task.ID)
task.Status = models.TaskStatusPending
task.Speed = 0
task.Chunk = models.GetChunksByTaskId(h.db, task.ID)
models.UpdateTask(h.db, &task)
lastTotalDownloaded := task.TotalDownloaded
go func() {
h.processDownload(es, &task, lastTotalDownloaded)
}()
}
c.JSON(http.StatusOK, gin.H{"ids": ids})
}
重新下载实现步骤:
- 获取所有 id
- 根据 id 获取所有任务
- 枚举所有任务,重置下载进度,根据 key 创建对应的 eventSource,更新任务状态,异步调用 processDownload 方法执行下载
- 等待下载结束
// Re download
func (h *Handlers) handleRestart(c *gin.Context) {
var ids []string
err := c.ShouldBindJSON(&ids)
if err != nil {
carrot.AbortWithJSONError(c, http.StatusBadRequest, err)
return
}
tasks, err := models.GetTaskByIds(h.db, ids)
if err != nil {
carrot.AbortWithJSONError(c, http.StatusInternalServerError, err)
return
}
for _, task := range tasks {
if task.Status != models.TaskStatusDownloaded {
carrot.AbortWithJSONError(c, http.StatusBadRequest, models.ErrStatusNotDownloaded)
return
}
es := h.createEventSourceWithKey(task.ID)
task.Status = models.TaskStatusPending
task.TotalDownloaded = 0
task.Progress = 0
task.Speed = 0
task.Chunk = make([]models.Chunk, task.ChunkNum)
models.UpdateTask(h.db, &task)
go func() {
h.processDownload(es, &task, 0)
}()
}
c.JSON(http.StatusOK, gin.H{"ids": ids})
}
限速下载实现步骤:
- 安装 `golang.org/x/time/rate` 库
- 创建一个限速器,设置每秒允许的请求数和桶的大小(这里限速单位是 MB/s,需要先将 MB/s 转换为 Bytes/s)
// Create a rate limiter maxDownloadSpeedInBytes := maxDownloadSpeed * 1000 * 1000 limiter := rate.NewLimiter(rate.Limit(maxDownloadSpeedInBytes), int(maxDownloadSpeedInBytes*0.1)) - 集成到 `downloadChunk`方法中
// Create a rate limiter
maxDownloadSpeedInBytes := maxDownloadSpeed * 1000 * 1000
limiter := rate.NewLimiter(rate.Limit(maxDownloadSpeedInBytes), int(maxDownloadSpeedInBytes))
for {
n, err := resp.Body.Read(buf)
if err != nil && err != io.EOF {
h.mu.Unlock()
return err
}
if n == 0 {
break
}
err = limiter.WaitN(es.ctx, n)
if err != nil {
h.mu.Unlock()
return err
}
_, err = outputFile.Write(buf[:n])
if err != nil {
h.mu.Unlock()
return err
}
task.TotalDownloaded += int64(n)
speed, progress, remainingTime := h.calculateDownloadData(task, startTime, lastTotalDownloaded)
es.Emit(DownloadProgress{
ID: task.ID,
Name: task.Name,
Progress: progress,
Speed: speed,
RemainingTime: remainingTime,
Status: task.Status,
})
}
1400

被折叠的 条评论
为什么被折叠?



