| |
|
|
| const fs = require('fs'); |
| const ts = require('typescript'); |
| const { tsquery } = require('@phenomnomnominal/tsquery'); |
|
|
| let data = ''; |
|
|
| process.stdin.on('data', function (chunk) { |
| data += chunk; |
| }); |
|
|
| process.stdin.on('end', function () { |
| const sourceFile = ts.createSourceFile('temp.ts', data.toString(), ts.ScriptTarget.ES2015, true); |
|
|
| |
| const semanticChunkNodeTypes = [ |
| 'FunctionDeclaration', |
| 'ArrowFunction', |
| 'ClassDeclaration', |
| 'InterfaceDeclaration', |
| 'EnumDeclaration', |
| 'TypeAliasDeclaration', |
| 'MethodDeclaration', |
| ]; |
|
|
| const semanticChunks = semanticChunkNodeTypes.flatMap(nodeType => |
| tsquery(sourceFile, nodeType) |
| ); |
|
|
| const jsonl = semanticChunks.map(chunk => { |
| const comments = ts.getLeadingCommentRanges(sourceFile.getFullText(), chunk.getFullStart()) || []; |
| const commentTexts = comments.map(comment => sourceFile.getFullText().slice(comment.pos, comment.end)).join('\n'); |
|
|
| |
| const contentWithComments = commentTexts + '\n' + chunk.getText(sourceFile); |
|
|
| return JSON.stringify({ |
| type: ts.SyntaxKind[chunk.kind], |
| content: contentWithComments |
| }); |
| }).join('\n'); |
|
|
| fs.writeFileSync('semantic_chunks.jsonl', jsonl); |
| }); |
|
|