Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
161 changes: 97 additions & 64 deletions src/export/csv.test.ts
Original file line number Diff line number Diff line change
@@ -1,13 +1,12 @@
import { describe, it, expect, vi, beforeEach } from 'vitest'
import { exportTableToCsvRoute } from './csv'
import { getTableData, createExportResponse } from './index'
import { executeOperation } from '.'
import { createResponse } from '../utils'
import type { DataSource } from '../types'
import type { StarbaseDBConfiguration } from '../handler'

vi.mock('./index', () => ({
getTableData: vi.fn(),
createExportResponse: vi.fn(),
vi.mock('.', () => ({
executeOperation: vi.fn(),
}))

vi.mock('../utils', () => ({
Expand Down Expand Up @@ -42,51 +41,43 @@ beforeEach(() => {
})

describe('CSV Export Module', () => {
it('should return a CSV file when table data exists', async () => {
vi.mocked(getTableData).mockResolvedValue([
{ id: 1, name: 'Alice', age: 30 },
{ id: 2, name: 'Bob', age: 25 },
])

vi.mocked(createExportResponse).mockReturnValue(
new Response('mocked-csv-content', {
headers: { 'Content-Type': 'text/csv' },
})
)
it('should stream a CSV file when table data exists', async () => {
vi.mocked(executeOperation)
.mockResolvedValueOnce([{ name: 'users' }])
.mockResolvedValueOnce([
{ name: 'id', pk: 1 },
{ name: 'name', pk: 0 },
{ name: 'age', pk: 0 },
])
.mockResolvedValueOnce([
{ id: 1, name: 'Alice', age: 30 },
{ id: 2, name: 'Bob', age: 25 },
])

const response = await exportTableToCsvRoute(
'users',
mockDataSource,
mockConfig
)

expect(getTableData).toHaveBeenCalledWith(
'users',
mockDataSource,
mockConfig
expect(response.headers.get('Content-Type')).toBe('text/csv')
expect(response.headers.get('Content-Disposition')).toBe(
'attachment; filename="users_export.csv"'
)
expect(createExportResponse).toHaveBeenCalledWith(
'id,name,age\n1,Alice,30\n2,Bob,25\n',
'users_export.csv',
'text/csv'
await expect(response.text()).resolves.toBe(
'id,name,age\n1,Alice,30\n2,Bob,25\n'
)
expect(response.headers.get('Content-Type')).toBe('text/csv')
})

it('should return 404 if table does not exist', async () => {
vi.mocked(getTableData).mockResolvedValue(null)
vi.mocked(executeOperation).mockResolvedValueOnce([])

const response = await exportTableToCsvRoute(
'non_existent_table',
mockDataSource,
mockConfig
)

expect(getTableData).toHaveBeenCalledWith(
'non_existent_table',
mockDataSource,
mockConfig
)
expect(response.status).toBe(404)

const jsonResponse: { error: string } = await response.json()
Expand All @@ -95,64 +86,105 @@ describe('CSV Export Module', () => {
)
})

it('should handle empty table (return only headers)', async () => {
vi.mocked(getTableData).mockResolvedValue([])

vi.mocked(createExportResponse).mockReturnValue(
new Response('mocked-csv-content', {
headers: { 'Content-Type': 'text/csv' },
})
)
it('should include headers for empty tables', async () => {
vi.mocked(executeOperation)
.mockResolvedValueOnce([{ name: 'empty_table' }])
.mockResolvedValueOnce([
{ name: 'id', pk: 1 },
{ name: 'name', pk: 0 },
])
.mockResolvedValueOnce([])

const response = await exportTableToCsvRoute(
'empty_table',
mockDataSource,
mockConfig
)

expect(getTableData).toHaveBeenCalledWith(
'empty_table',
await expect(response.text()).resolves.toBe('id,name\n')
})

it('should escape commas, quotes, and newlines in CSV values', async () => {
vi.mocked(executeOperation)
.mockResolvedValueOnce([{ name: 'special_chars' }])
.mockResolvedValueOnce([
{ name: 'id', pk: 1 },
{ name: 'name', pk: 0 },
{ name: 'bio', pk: 0 },
])
.mockResolvedValueOnce([
{
id: 1,
name: 'Sahithi, is',
bio: 'my forever "penguin"\nline',
},
])

const response = await exportTableToCsvRoute(
'special_chars',
mockDataSource,
mockConfig
)
expect(createExportResponse).toHaveBeenCalledWith(
'',
'empty_table_export.csv',
'text/csv'

await expect(response.text()).resolves.toBe(
'id,name,bio\n1,"Sahithi, is","my forever ""penguin""\nline"\n'
)
expect(response.headers.get('Content-Type')).toBe('text/csv')
})

it('should escape commas and quotes in CSV values', async () => {
vi.mocked(getTableData).mockResolvedValue([
{ id: 1, name: 'Sahithi, is', bio: 'my forever "penguin"' },
])

vi.mocked(createExportResponse).mockReturnValue(
new Response('mocked-csv-content', {
headers: { 'Content-Type': 'text/csv' },
})
)
it('should page table data with a stable order instead of loading the full table', async () => {
const firstPage = Array.from({ length: 500 }, (_, index) => ({
id: index + 1,
name: `User ${index + 1}`,
}))

vi.mocked(executeOperation)
.mockResolvedValueOnce([{ name: 'users' }])
.mockResolvedValueOnce([
{ name: 'id', pk: 1 },
{ name: 'name', pk: 0 },
])
.mockResolvedValueOnce(firstPage)
.mockResolvedValueOnce([{ id: 501, name: 'Last User' }])

const response = await exportTableToCsvRoute(
'special_chars',
'users',
mockDataSource,
mockConfig
)

expect(createExportResponse).toHaveBeenCalledWith(
'id,name,bio\n1,"Sahithi, is","my forever ""penguin"""\n',
'special_chars_export.csv',
'text/csv'
const csv = await response.text()

expect(csv).toContain('501,Last User\n')
expect(executeOperation).toHaveBeenNthCalledWith(
3,
[
{
sql: 'SELECT * FROM "users" ORDER BY "id" LIMIT ? OFFSET ?;',
params: [500, 0],
},
],
mockDataSource,
mockConfig
)
expect(executeOperation).toHaveBeenNthCalledWith(
4,
[
{
sql: 'SELECT * FROM "users" ORDER BY "id" LIMIT ? OFFSET ?;',
params: [500, 500],
},
],
mockDataSource,
mockConfig
)
expect(response.headers.get('Content-Type')).toBe('text/csv')
})

it('should return 500 on an unexpected error', async () => {
it('should return 500 on an unexpected error before streaming starts', async () => {
const consoleErrorMock = vi
.spyOn(console, 'error')
.mockImplementation(() => {})
vi.mocked(getTableData).mockRejectedValue(new Error('Database Error'))
vi.mocked(executeOperation).mockRejectedValue(
new Error('Database Error')
)

const response = await exportTableToCsvRoute(
'users',
Expand All @@ -163,5 +195,6 @@ describe('CSV Export Module', () => {
expect(response.status).toBe(500)
const jsonResponse: { error: string } = await response.json()
expect(jsonResponse.error).toBe('Failed to export table to CSV')
consoleErrorMock.mockRestore()
})
})
62 changes: 31 additions & 31 deletions src/export/csv.ts
Original file line number Diff line number Diff line change
@@ -1,51 +1,51 @@
import { getTableData, createExportResponse } from './index'
import { createResponse } from '../utils'
import { DataSource } from '../types'
import { StarbaseDBConfiguration } from '../handler'
import {
createStreamingExportResponse,
formatCsvValue,
getTablePagePlan,
iterateTableRows,
tableExists,
} from './streaming'

async function* csvTableChunks(
tableName: string,
dataSource: DataSource,
config: StarbaseDBConfiguration
): AsyncGenerator<string> {
const pagePlan = await getTablePagePlan(tableName, dataSource, config)

if (pagePlan.columns.length) {
yield `${pagePlan.columns.map(formatCsvValue).join(',')}\n`
}

for await (const row of iterateTableRows(
tableName,
dataSource,
config,
pagePlan
)) {
yield `${pagePlan.columns.map((column) => formatCsvValue(row[column])).join(',')}\n`
}
}

export async function exportTableToCsvRoute(
tableName: string,
dataSource: DataSource,
config: StarbaseDBConfiguration
): Promise<Response> {
try {
const data = await getTableData(tableName, dataSource, config)

if (data === null) {
if (!(await tableExists(tableName, dataSource, config))) {
return createResponse(
undefined,
`Table '${tableName}' does not exist.`,
404
)
}

// Convert the result to CSV
let csvContent = ''
if (data.length > 0) {
// Add headers
csvContent += Object.keys(data[0]).join(',') + '\n'

// Add data rows
data.forEach((row: any) => {
csvContent +=
Object.values(row)
.map((value) => {
if (
typeof value === 'string' &&
(value.includes(',') ||
value.includes('"') ||
value.includes('\n'))
) {
return `"${value.replace(/"/g, '""')}"`
}
return value
})
.join(',') + '\n'
})
}

return createExportResponse(
csvContent,
return createStreamingExportResponse(
csvTableChunks(tableName, dataSource, config),
`${tableName}_export.csv`,
'text/csv'
)
Expand Down
Loading