In modern web applications, handling large file uploads is a common requirement. Whether you're building a file sharing service, a media platform, or a document management system, you'll likely face challenges with uploading large files. Traditional file upload methods can be problematic, leading to timeouts, memory issues, and poor user experience. In this post, we'll explore how to implement a robust chunk-based file upload system using Go and JavaScript.
Traditional file uploads work well for small files, but they break down when dealing with larger files. Some common issues include:
Chunk-based uploading solves these problems by breaking large files into smaller, manageable pieces and uploading them sequentially.
The core idea behind chunk-based uploads is simple: instead of sending the entire file at once, we split it into smaller chunks, upload each chunk separately, and reassemble them on the server. This approach offers several benefits:
Let's see how to implement this in practice.
Our implementation consists of two main parts:
Let's dive into each part.
First, let's look at how we handle file chunking in the browser. We'll set a chunk size of 512KB:
const CHUNK_SIZE = 512 * 1000; // 512KB chunks function handleFiles() { selectedFile = this.files[0]; let fullChunks = Math.floor(selectedFile.size / CHUNK_SIZE); // Update UI with file information const fileSizeEl = document.getElementById("file-size"); fileSizeEl.innerHTML = "File size:" + selectedFile.size; const chunkCountEl = document.getElementById("chunk-count"); chunkCountEl.innerHTML = "Count count:" + fullChunks; }
For each file upload, we generate a unique identifier using UUID to track all chunks belonging to the same file:
function generateUUID() { let d = new Date().getTime(); let d2 = (typeof performance !== "undefined" && performance.now && performance.now() * 1000) || 0; return "xxxxxxxx-xxxx-4xxx-yxxx-xxxxxxxxxxxx".replace(/[xy]/g, function (c) { let r = Math.random() * 16; if (d > 0) { r = (d + r) % 16 | 0; d = Math.floor(d / 16); } else { r = (d2 + r) % 16 | 0; d2 = Math.floor(d2 / 16); } return (c === "x" ? r : (r & 0x3) | 0x8).toString(16); }); }
The upload process handles both single files and chunked uploads based on file size:
async function handleClickUpload() { try { const fileId = generateUUID(); const fileSize = selectedFile.size; const fileName = selectedFile.name; const fullChunks = Math.floor(fileSize / CHUNK_SIZE); if (fullChunks > 0) { // Handle chunked upload for (let i = 0; i < fullChunks; i++) { const offset = CHUNK_SIZE * i; const limit = CHUNK_SIZE * (i + 1); const metadata = { order: i, fileId, offset, limit, fileSize, fileName, }; const chunkedFile = selectedFile.slice(offset, limit); data.append("file", chunkedFile); data.append("metadata", JSON.stringify(metadata)); const res = await fetch(chunkUploadURL, { method: "POST", body: data, }); if (!res.ok) { throw new Error("Response status:" + res.status); } const json = await res.json(); const fileUploadedCount = document.getElementById("uploaded-count"); uploadCount++; fileUploadedCount.innerHTML = "File uploaded count:" + uploadCount; } if (remainedChunk > 0) { const data = new FormData(); const offset = fileSize - remainedChunk; const limit = fileSize; const metadata = { order: fullChunks, fileId, offset, limit, fileSize, fileName, }; const chunkedFile = selectedFile.slice(offset, limit); data.append("file", chunkedFile); data.append("metadata", JSON.stringify(metadata)); const res = await fetch(chunkUploadURL, { method: "POST", body: data, }); if (!res.oke) { throw new Error("Response status:" + res.status); } const json = await res.json(); } } else { // Handle single file upload // ... } } catch (err) { console.error("error click upload", err); } }
Our Go backend uses the Gin framework to handle the uploads. Here's the core structure:
type Metadata struct { Order int `json:"order"` FileId string `json:"fileId"` Offset int `json:"offset"` Limit int `json:"limit"` FileSize int `json:"fileSize"` FileName string `json:"fileName"` }
The server handles two types of uploads:
For chunked uploads, we store each chunk temporarily and reassemble them when all chunks are received:
r.POST("/split-upload", func(c *gin.Context) { // Handle file chunk and metadata // ... if metadata.FileSize == metadata.Limit { // All chunks received, start reassembly chunks, err := filepath.Glob(filepath.Join("./uploads/temp", fmt.Sprintf("*_%s", metadata.FileId))) // Sort chunks by order sort.Slice(chunks, func(i, j int) bool { orderI, _ := strconv.Atoi(string(filepath.Base(chunks[i])[0])) orderJ, _ := strconv.Atoi(string(filepath.Base(chunks[j])[0])) return orderI < orderJ }) // Merge chunks into final file finalPath := filepath.Join("./uploads", fmt.Sprintf("merged_%s", metadata.FileName)) finalFile, err := os.Create(finalPath) if err != nil { c.String(http.StatusBadRequest, "error merging file: %s", err.Error()) return } defer finalFile.Close() for _, chunk := range chunks { chunkFile, err := os.Open(chunk) if err != nil { c.String(http.StatusBadRequest, "error open chunk file: %s", err.Error()) return } _, err = io.Copy(finalFile, chunkFile) chunkFile.Close() if err != nil { c.String(http.StatusBadRequest, "error merging chunk file: %s", err.Error()) return } } // Cleanup temporary chunks // ... } })
When implementing chunk uploads, consider:
This implementation could be enhanced with:
Chunk-based file uploading provides a robust solution for handling large files in web applications. While the implementation requires more complexity than traditional uploads, the benefits in reliability and user experience make it worthwhile for applications dealing with large files.
The code demonstrated here provides a foundation that you can build upon based on your specific needs. Consider factors like your expected file sizes, user experience requirements, and server capabilities when adapting this solution.
Remember to properly test the implementation with various file sizes and types, and implement appropriate security measures before deploying to production.
You can read the code here in my repository
GitHub repo: halosatrio/split-upload