mirror of
https://github.com/kjanat/articulate-parser.git
synced 2026-01-16 19:42:12 +01:00
Compare commits
24 Commits
hotfix/two
...
dependabot
| Author | SHA1 | Date | |
|---|---|---|---|
| 02941138ee | |||
|
94a7924bed
|
|||
|
90b9d557d8
|
|||
|
33ff267644
|
|||
|
33673d661b
|
|||
| 41f3f5c4e2 | |||
|
d644094999
|
|||
|
71d1429048
|
|||
| bd308e4dfc | |||
| 227f88cb9b | |||
| b56c9fa29f | |||
| d8e4d97841 | |||
| fe588dadda | |||
| 68c6f4e408 | |||
| 37927a36b6 | |||
| e6977d3374 | |||
| 2790064ad5 | |||
| 65469ea52e | |||
| 2db2e0b1a3 | |||
| 6317ce268b | |||
| 59f2de9d22 | |||
| f8fecc3967 | |||
| af15bcccd4 | |||
| 422b56aa86 |
42
.dprint.jsonc
Normal file
42
.dprint.jsonc
Normal file
@ -0,0 +1,42 @@
|
|||||||
|
{
|
||||||
|
"typescript": {
|
||||||
|
},
|
||||||
|
"json": {
|
||||||
|
},
|
||||||
|
"markdown": {
|
||||||
|
},
|
||||||
|
"toml": {
|
||||||
|
},
|
||||||
|
"dockerfile": {
|
||||||
|
},
|
||||||
|
"oxc": {
|
||||||
|
},
|
||||||
|
"ruff": {
|
||||||
|
},
|
||||||
|
"jupyter": {
|
||||||
|
},
|
||||||
|
"malva": {
|
||||||
|
},
|
||||||
|
"markup": {
|
||||||
|
},
|
||||||
|
"yaml": {
|
||||||
|
},
|
||||||
|
"excludes": [
|
||||||
|
"**/node_modules",
|
||||||
|
"**/*-lock.json",
|
||||||
|
],
|
||||||
|
"plugins": [
|
||||||
|
"https://plugins.dprint.dev/typescript-0.95.13.wasm",
|
||||||
|
"https://plugins.dprint.dev/json-0.21.1.wasm",
|
||||||
|
"https://plugins.dprint.dev/markdown-0.20.0.wasm",
|
||||||
|
"https://plugins.dprint.dev/toml-0.7.0.wasm",
|
||||||
|
"https://plugins.dprint.dev/dockerfile-0.3.3.wasm",
|
||||||
|
"https://plugins.dprint.dev/oxc-0.1.0.wasm",
|
||||||
|
"https://plugins.dprint.dev/ruff-0.6.11.wasm",
|
||||||
|
"https://plugins.dprint.dev/jupyter-0.2.1.wasm",
|
||||||
|
"https://plugins.dprint.dev/g-plane/malva-v0.15.1.wasm",
|
||||||
|
"https://plugins.dprint.dev/g-plane/markup_fmt-v0.25.3.wasm",
|
||||||
|
"https://plugins.dprint.dev/g-plane/pretty_yaml-v0.5.1.wasm",
|
||||||
|
"https://plugins.dprint.dev/exec-0.6.0.json@a054130d458f124f9b5c91484833828950723a5af3f8ff2bd1523bd47b83b364",
|
||||||
|
],
|
||||||
|
}
|
||||||
20
.github/CODE_OF_CONDUCT.md
vendored
20
.github/CODE_OF_CONDUCT.md
vendored
@ -17,23 +17,23 @@ diverse, inclusive, and healthy community.
|
|||||||
Examples of behavior that contributes to a positive environment for our
|
Examples of behavior that contributes to a positive environment for our
|
||||||
community include:
|
community include:
|
||||||
|
|
||||||
- Demonstrating empathy and kindness toward other people
|
- Demonstrating empathy and kindness toward other people
|
||||||
- Being respectful of differing opinions, viewpoints, and experiences
|
- Being respectful of differing opinions, viewpoints, and experiences
|
||||||
- Giving and gracefully accepting constructive feedback
|
- Giving and gracefully accepting constructive feedback
|
||||||
- Accepting responsibility and apologizing to those affected by our mistakes,
|
- Accepting responsibility and apologizing to those affected by our mistakes,
|
||||||
and learning from the experience
|
and learning from the experience
|
||||||
- Focusing on what is best not just for us as individuals, but for the
|
- Focusing on what is best not just for us as individuals, but for the
|
||||||
overall community
|
overall community
|
||||||
|
|
||||||
Examples of unacceptable behavior include:
|
Examples of unacceptable behavior include:
|
||||||
|
|
||||||
- The use of sexualized language or imagery, and sexual attention or
|
- The use of sexualized language or imagery, and sexual attention or
|
||||||
advances of any kind
|
advances of any kind
|
||||||
- Trolling, insulting or derogatory comments, and personal or political attacks
|
- Trolling, insulting or derogatory comments, and personal or political attacks
|
||||||
- Public or private harassment
|
- Public or private harassment
|
||||||
- Publishing others' private information, such as a physical or email
|
- Publishing others' private information, such as a physical or email
|
||||||
address, without their explicit permission
|
address, without their explicit permission
|
||||||
- Other conduct which could reasonably be considered inappropriate in a
|
- Other conduct which could reasonably be considered inappropriate in a
|
||||||
professional setting
|
professional setting
|
||||||
|
|
||||||
## Enforcement Responsibilities
|
## Enforcement Responsibilities
|
||||||
|
|||||||
130
.github/CONTRIBUTING.md
vendored
130
.github/CONTRIBUTING.md
vendored
@ -12,73 +12,73 @@ This project and everyone participating in it is governed by our Code of Conduct
|
|||||||
|
|
||||||
Before creating bug reports, please check existing issues as you might find that the issue has already been reported. When creating a bug report, include as many details as possible:
|
Before creating bug reports, please check existing issues as you might find that the issue has already been reported. When creating a bug report, include as many details as possible:
|
||||||
|
|
||||||
- Use the bug report template
|
- Use the bug report template
|
||||||
- Include sample Articulate Rise content that reproduces the issue
|
- Include sample Articulate Rise content that reproduces the issue
|
||||||
- Provide your environment details (OS, Go version, etc.)
|
- Provide your environment details (OS, Go version, etc.)
|
||||||
- Include error messages and stack traces
|
- Include error messages and stack traces
|
||||||
|
|
||||||
### Suggesting Enhancements
|
### Suggesting Enhancements
|
||||||
|
|
||||||
Enhancement suggestions are welcome! Please use the feature request template and include:
|
Enhancement suggestions are welcome! Please use the feature request template and include:
|
||||||
|
|
||||||
- A clear description of the enhancement
|
- A clear description of the enhancement
|
||||||
- Your use case and why this would be valuable
|
- Your use case and why this would be valuable
|
||||||
- Any implementation ideas you might have
|
- Any implementation ideas you might have
|
||||||
|
|
||||||
### Pull Requests
|
### Pull Requests
|
||||||
|
|
||||||
1. **Fork the repository** and create your branch from `master`
|
1. **Fork the repository** and create your branch from `master`
|
||||||
2. **Make your changes** following our coding standards
|
2. **Make your changes** following our coding standards
|
||||||
3. **Add tests** for any new functionality
|
3. **Add tests** for any new functionality
|
||||||
4. **Ensure all tests pass** by running `go test ./...`
|
4. **Ensure all tests pass** by running `go test ./...`
|
||||||
5. **Run `go fmt`** to format your code
|
5. **Run `go fmt`** to format your code
|
||||||
6. **Run `go vet`** to check for common issues
|
6. **Run `go vet`** to check for common issues
|
||||||
7. **Update documentation** if needed
|
7. **Update documentation** if needed
|
||||||
8. **Create a pull request** with a clear title and description
|
8. **Create a pull request** with a clear title and description
|
||||||
|
|
||||||
## Development Setup
|
## Development Setup
|
||||||
|
|
||||||
1. **Prerequisites:**
|
1. **Prerequisites:**
|
||||||
|
|
||||||
- Go 1.21 or later
|
- Go 1.21 or later
|
||||||
- Git
|
- Git
|
||||||
|
|
||||||
2. **Clone and setup:**
|
2. **Clone and setup:**
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
git clone https://github.com/your-username/articulate-parser.git
|
git clone https://github.com/your-username/articulate-parser.git
|
||||||
cd articulate-parser
|
cd articulate-parser
|
||||||
go mod download
|
go mod download
|
||||||
```
|
```
|
||||||
|
|
||||||
3. **Run tests:**
|
3. **Run tests:**
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
go test -v ./...
|
go test -v ./...
|
||||||
```
|
```
|
||||||
|
|
||||||
4. **Build:**
|
4. **Build:**
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
go build main.go
|
go build main.go
|
||||||
```
|
```
|
||||||
|
|
||||||
## Coding Standards
|
## Coding Standards
|
||||||
|
|
||||||
### Go Style Guide
|
### Go Style Guide
|
||||||
|
|
||||||
- Follow the [Go Code Review Comments](https://github.com/golang/go/wiki/CodeReviewComments)
|
- Follow the [Go Code Review Comments](https://github.com/golang/go/wiki/CodeReviewComments)
|
||||||
- Use `gofmt` to format your code
|
- Use `gofmt` to format your code
|
||||||
- Use meaningful variable and function names
|
- Use meaningful variable and function names
|
||||||
- Add comments for exported functions and types
|
- Add comments for exported functions and types
|
||||||
- Keep functions focused and small
|
- Keep functions focused and small
|
||||||
|
|
||||||
### Testing
|
### Testing
|
||||||
|
|
||||||
- Write tests for new functionality
|
- Write tests for new functionality
|
||||||
- Use table-driven tests where appropriate
|
- Use table-driven tests where appropriate
|
||||||
- Aim for good test coverage
|
- Aim for good test coverage
|
||||||
- Test error cases and edge conditions
|
- Test error cases and edge conditions
|
||||||
|
|
||||||
### Commit Messages
|
### Commit Messages
|
||||||
|
|
||||||
@ -112,19 +112,19 @@ articulate-parser/
|
|||||||
|
|
||||||
### New Content Types
|
### New Content Types
|
||||||
|
|
||||||
1. Add the content type definition to `types/`
|
1. Add the content type definition to `types/`
|
||||||
2. Implement parsing logic in `parser/`
|
2. Implement parsing logic in `parser/`
|
||||||
3. Add export handling in `exporters/`
|
3. Add export handling in `exporters/`
|
||||||
4. Write comprehensive tests
|
4. Write comprehensive tests
|
||||||
5. Update documentation
|
5. Update documentation
|
||||||
|
|
||||||
### New Export Formats
|
### New Export Formats
|
||||||
|
|
||||||
1. Create a new exporter in `exporters/`
|
1. Create a new exporter in `exporters/`
|
||||||
2. Implement the `Exporter` interface
|
2. Implement the `Exporter` interface
|
||||||
3. Add CLI support in `main.go`
|
3. Add CLI support in `main.go`
|
||||||
4. Add tests with sample output
|
4. Add tests with sample output
|
||||||
5. Update README with usage examples
|
5. Update README with usage examples
|
||||||
|
|
||||||
## Testing
|
## Testing
|
||||||
|
|
||||||
@ -146,31 +146,31 @@ go test -run TestSpecificFunction ./...
|
|||||||
|
|
||||||
### Test Data
|
### Test Data
|
||||||
|
|
||||||
- Add sample Articulate Rise JSON files to `tests/data/`
|
- Add sample Articulate Rise JSON files to `tests/data/`
|
||||||
- Include both simple and complex content examples
|
- Include both simple and complex content examples
|
||||||
- Test edge cases and error conditions
|
- Test edge cases and error conditions
|
||||||
|
|
||||||
## Documentation
|
## Documentation
|
||||||
|
|
||||||
- Update the README for user-facing changes
|
- Update the README for user-facing changes
|
||||||
- Add inline code comments for complex logic
|
- Add inline code comments for complex logic
|
||||||
- Update examples when adding new features
|
- Update examples when adding new features
|
||||||
- Keep the feature list current
|
- Keep the feature list current
|
||||||
|
|
||||||
## Release Process
|
## Release Process
|
||||||
|
|
||||||
Releases are handled by maintainers:
|
Releases are handled by maintainers:
|
||||||
|
|
||||||
1. Version bumping follows semantic versioning
|
1. Version bumping follows semantic versioning
|
||||||
2. Releases are created from the `master` branch
|
2. Releases are created from the `master` branch
|
||||||
3. GitHub Actions automatically builds and publishes releases
|
3. GitHub Actions automatically builds and publishes releases
|
||||||
4. Release notes are auto-generated from commits
|
4. Release notes are auto-generated from commits
|
||||||
|
|
||||||
## Questions?
|
## Questions?
|
||||||
|
|
||||||
- Open a discussion for general questions
|
- Open a discussion for general questions
|
||||||
- Use the question issue template for specific help
|
- Use the question issue template for specific help
|
||||||
- Check existing issues and documentation first
|
- Check existing issues and documentation first
|
||||||
|
|
||||||
## Recognition
|
## Recognition
|
||||||
|
|
||||||
|
|||||||
4
.github/ISSUE_TEMPLATE/bug_report.yml
vendored
4
.github/ISSUE_TEMPLATE/bug_report.yml
vendored
@ -1,7 +1,7 @@
|
|||||||
name: Bug Report
|
name: Bug Report
|
||||||
description: Create a report to help us improve
|
description: Create a report to help us improve
|
||||||
title: '[BUG] '
|
title: "[BUG] "
|
||||||
labels: ['bug', 'triage']
|
labels: ["bug", "triage"]
|
||||||
body:
|
body:
|
||||||
- type: markdown
|
- type: markdown
|
||||||
attributes:
|
attributes:
|
||||||
|
|||||||
36
.github/PULL_REQUEST_TEMPLATE.md
vendored
36
.github/PULL_REQUEST_TEMPLATE.md
vendored
@ -5,31 +5,34 @@
|
|||||||
## Related Issue
|
## Related Issue
|
||||||
|
|
||||||
<!-- Link to the issue this PR addresses using the syntax: Fixes #issue_number -->
|
<!-- Link to the issue this PR addresses using the syntax: Fixes #issue_number -->
|
||||||
|
|
||||||
Fixes #
|
Fixes #
|
||||||
|
|
||||||
## Type of Change
|
## Type of Change
|
||||||
|
|
||||||
<!-- Mark the appropriate option with an "x" -->
|
<!-- Mark the appropriate option with an "x" -->
|
||||||
- [ ] Bug fix (non-breaking change which fixes an issue)
|
|
||||||
- [ ] New feature (non-breaking change which adds functionality)
|
- [ ] Bug fix (non-breaking change which fixes an issue)
|
||||||
- [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
|
- [ ] New feature (non-breaking change which adds functionality)
|
||||||
- [ ] Documentation update
|
- [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
|
||||||
- [ ] Performance improvement
|
- [ ] Documentation update
|
||||||
- [ ] Code refactoring (no functional changes)
|
- [ ] Performance improvement
|
||||||
- [ ] Test updates
|
- [ ] Code refactoring (no functional changes)
|
||||||
|
- [ ] Test updates
|
||||||
|
|
||||||
## Checklist
|
## Checklist
|
||||||
|
|
||||||
<!-- Mark the items you've completed with an "x" -->
|
<!-- Mark the items you've completed with an "x" -->
|
||||||
- [ ] My code follows the style guidelines of this project
|
|
||||||
- [ ] I have performed a self-review of my code
|
- [ ] My code follows the style guidelines of this project
|
||||||
- [ ] I have added comments to complex logic
|
- [ ] I have performed a self-review of my code
|
||||||
- [ ] I have updated the documentation
|
- [ ] I have added comments to complex logic
|
||||||
- [ ] I have added tests that prove my fix is effective or that my feature works
|
- [ ] I have updated the documentation
|
||||||
- [ ] New and existing unit tests pass locally with my changes
|
- [ ] I have added tests that prove my fix is effective or that my feature works
|
||||||
- [ ] I have checked for potential breaking changes
|
- [ ] New and existing unit tests pass locally with my changes
|
||||||
- [ ] No new warnings are generated
|
- [ ] I have checked for potential breaking changes
|
||||||
- [ ] The commit message follows our guidelines
|
- [ ] No new warnings are generated
|
||||||
|
- [ ] The commit message follows our guidelines
|
||||||
|
|
||||||
## Screenshots (if appropriate)
|
## Screenshots (if appropriate)
|
||||||
|
|
||||||
@ -42,6 +45,7 @@ Fixes #
|
|||||||
## Testing Instructions
|
## Testing Instructions
|
||||||
|
|
||||||
<!-- Provide steps to test the changes, if applicable -->
|
<!-- Provide steps to test the changes, if applicable -->
|
||||||
|
|
||||||
1.
|
1.
|
||||||
2.
|
2.
|
||||||
3.
|
3.
|
||||||
|
|||||||
32
.github/SECURITY.md
vendored
32
.github/SECURITY.md
vendored
@ -13,32 +13,32 @@ Currently, the following versions of Articulate Rise Parser are supported with s
|
|||||||
|
|
||||||
We take the security of Articulate Rise Parser seriously. If you believe you have found a security vulnerability, please follow these steps:
|
We take the security of Articulate Rise Parser seriously. If you believe you have found a security vulnerability, please follow these steps:
|
||||||
|
|
||||||
1. **Do not disclose the vulnerability publicly** - Please do not create a public GitHub issue for security vulnerabilities.
|
1. **Do not disclose the vulnerability publicly** - Please do not create a public GitHub issue for security vulnerabilities.
|
||||||
2. **Email the details to [security+articulate-parser@kjanat.com]** - Include as much information as possible about the vulnerability.
|
2. **Email the details to [security+articulate-parser@kjanat.com]** - Include as much information as possible about the vulnerability.
|
||||||
3. **Wait for a response** - We will acknowledge your email within 48 hours and provide an estimated timeline for a fix.
|
3. **Wait for a response** - We will acknowledge your email within 48 hours and provide an estimated timeline for a fix.
|
||||||
4. **Work with us** - We may ask for additional information to help us understand and address the issue.
|
4. **Work with us** - We may ask for additional information to help us understand and address the issue.
|
||||||
|
|
||||||
## What to Include in a Report
|
## What to Include in a Report
|
||||||
|
|
||||||
When reporting a vulnerability, please include:
|
When reporting a vulnerability, please include:
|
||||||
|
|
||||||
- A clear description of the issue
|
- A clear description of the issue
|
||||||
- Steps to reproduce the vulnerability
|
- Steps to reproduce the vulnerability
|
||||||
- The potential impact of the vulnerability
|
- The potential impact of the vulnerability
|
||||||
- Any possible mitigations you've identified
|
- Any possible mitigations you've identified
|
||||||
|
|
||||||
## What to Expect
|
## What to Expect
|
||||||
|
|
||||||
- We will acknowledge receipt of your vulnerability report within 48 hours.
|
- We will acknowledge receipt of your vulnerability report within 48 hours.
|
||||||
- We will provide regular updates about our progress.
|
- We will provide regular updates about our progress.
|
||||||
- We will notify you when the vulnerability is fixed.
|
- We will notify you when the vulnerability is fixed.
|
||||||
- With your permission, we will include your name in the acknowledgments.
|
- With your permission, we will include your name in the acknowledgments.
|
||||||
|
|
||||||
## Security Measures
|
## Security Measures
|
||||||
|
|
||||||
This project follows these security practices:
|
This project follows these security practices:
|
||||||
|
|
||||||
- Regular dependency updates via Dependabot
|
- Regular dependency updates via Dependabot
|
||||||
- CodeQL security scanning
|
- CodeQL security scanning
|
||||||
- Automated testing for each pull request
|
- Automated testing for each pull request
|
||||||
- Code review requirements for all changes
|
- Code review requirements for all changes
|
||||||
|
|||||||
110
.github/dependabot.yml
vendored
110
.github/dependabot.yml
vendored
@ -1,86 +1,86 @@
|
|||||||
version: 2
|
version: 2
|
||||||
updates:
|
updates:
|
||||||
# Check for updates to GitHub Actions
|
# Check for updates to GitHub Actions
|
||||||
- package-ecosystem: 'github-actions'
|
- package-ecosystem: "github-actions"
|
||||||
directory: '/'
|
directory: "/"
|
||||||
schedule:
|
schedule:
|
||||||
interval: 'weekly'
|
interval: "weekly"
|
||||||
day: 'monday'
|
day: "monday"
|
||||||
time: '07:00'
|
time: "07:00"
|
||||||
timezone: 'Europe/Amsterdam'
|
timezone: "Europe/Amsterdam"
|
||||||
open-pull-requests-limit: 10
|
open-pull-requests-limit: 2
|
||||||
labels:
|
labels:
|
||||||
- 'dependencies'
|
- "dependencies"
|
||||||
- 'dependencies/github-actions'
|
- "dependencies/github-actions"
|
||||||
commit-message:
|
commit-message:
|
||||||
prefix: 'ci'
|
prefix: "ci"
|
||||||
include: 'scope'
|
include: "scope"
|
||||||
|
|
||||||
# Check for updates to Docker
|
# Check for updates to Docker
|
||||||
- package-ecosystem: 'docker'
|
- package-ecosystem: "docker"
|
||||||
directory: '/'
|
directory: "/"
|
||||||
schedule:
|
schedule:
|
||||||
interval: 'weekly'
|
interval: "weekly"
|
||||||
day: 'monday'
|
day: "monday"
|
||||||
time: '07:00'
|
time: "07:00"
|
||||||
timezone: 'Europe/Amsterdam'
|
timezone: "Europe/Amsterdam"
|
||||||
open-pull-requests-limit: 10
|
open-pull-requests-limit: 2
|
||||||
labels:
|
labels:
|
||||||
- 'dependencies'
|
- "dependencies"
|
||||||
- 'dependencies/docker'
|
- "dependencies/docker"
|
||||||
commit-message:
|
commit-message:
|
||||||
prefix: 'docker'
|
prefix: "docker"
|
||||||
include: 'scope'
|
include: "scope"
|
||||||
groups:
|
groups:
|
||||||
docker-images:
|
docker:
|
||||||
patterns:
|
patterns:
|
||||||
- '*'
|
- "*"
|
||||||
update-types:
|
update-types:
|
||||||
- 'minor'
|
- "minor"
|
||||||
- 'patch'
|
- "patch"
|
||||||
|
|
||||||
# Check for updates to Docker Compose
|
# Check for updates to Docker Compose
|
||||||
- package-ecosystem: 'docker-compose'
|
- package-ecosystem: "docker-compose"
|
||||||
directory: '/'
|
directory: "/"
|
||||||
schedule:
|
schedule:
|
||||||
interval: 'weekly'
|
interval: "weekly"
|
||||||
day: 'monday'
|
day: "monday"
|
||||||
time: '07:00'
|
time: "07:00"
|
||||||
timezone: 'Europe/Amsterdam'
|
timezone: "Europe/Amsterdam"
|
||||||
open-pull-requests-limit: 10
|
open-pull-requests-limit: 2
|
||||||
labels:
|
labels:
|
||||||
- 'dependencies'
|
- "dependencies"
|
||||||
- 'dependencies/docker-compose'
|
- "dependencies/docker-compose"
|
||||||
commit-message:
|
commit-message:
|
||||||
prefix: 'docker'
|
prefix: "docker"
|
||||||
include: 'scope'
|
include: "scope"
|
||||||
groups:
|
groups:
|
||||||
docker-compose:
|
docker:
|
||||||
patterns:
|
patterns:
|
||||||
- '*'
|
- "*"
|
||||||
update-types:
|
update-types:
|
||||||
- 'minor'
|
- "minor"
|
||||||
- 'patch'
|
- "patch"
|
||||||
|
|
||||||
# Check for updates to Go modules
|
# Check for updates to Go modules
|
||||||
- package-ecosystem: 'gomod'
|
- package-ecosystem: "gomod"
|
||||||
directory: '/'
|
directory: "/"
|
||||||
schedule:
|
schedule:
|
||||||
interval: 'weekly'
|
interval: "weekly"
|
||||||
day: 'monday'
|
day: "monday"
|
||||||
time: '07:00'
|
time: "07:00"
|
||||||
timezone: 'Europe/Amsterdam'
|
timezone: "Europe/Amsterdam"
|
||||||
open-pull-requests-limit: 10
|
open-pull-requests-limit: 2
|
||||||
labels:
|
labels:
|
||||||
- 'dependencies'
|
- "dependencies"
|
||||||
- 'dependencies/go'
|
- "dependencies/go"
|
||||||
commit-message:
|
commit-message:
|
||||||
prefix: 'deps'
|
prefix: "deps"
|
||||||
include: 'scope'
|
include: "scope"
|
||||||
groups:
|
groups:
|
||||||
go-modules:
|
go-modules:
|
||||||
patterns:
|
patterns:
|
||||||
- '*'
|
- "*"
|
||||||
update-types:
|
update-types:
|
||||||
- 'minor'
|
- "minor"
|
||||||
- 'patch'
|
- "patch"
|
||||||
|
|||||||
44
.github/workflows/autofix.yml
vendored
44
.github/workflows/autofix.yml
vendored
@ -2,7 +2,7 @@ name: autofix.ci
|
|||||||
on:
|
on:
|
||||||
pull_request:
|
pull_request:
|
||||||
push:
|
push:
|
||||||
branches: [ "master" ]
|
branches: ["master"]
|
||||||
permissions:
|
permissions:
|
||||||
contents: read
|
contents: read
|
||||||
|
|
||||||
@ -10,16 +10,36 @@ jobs:
|
|||||||
autofix:
|
autofix:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- name: Checkout code
|
||||||
- uses: actions/setup-go@v5
|
uses: actions/checkout@v6
|
||||||
with:
|
|
||||||
go-version-file: 'go.mod'
|
|
||||||
|
|
||||||
# goimports works like gofmt, but also fixes imports.
|
- name: Install Task
|
||||||
# see https://pkg.go.dev/golang.org/x/tools/cmd/goimports
|
uses: go-task/setup-task@v1
|
||||||
- run: go install golang.org/x/tools/cmd/goimports@latest
|
|
||||||
- run: goimports -w .
|
|
||||||
# of course we can also do just this instead:
|
|
||||||
# - run: gofmt -w .
|
|
||||||
|
|
||||||
- uses: autofix-ci/action@551dded8c6cc8a1054039c8bc0b8b48c51dfc6ef
|
- uses: actions/setup-go@v6
|
||||||
|
with: { go-version-file: "go.mod" }
|
||||||
|
|
||||||
|
- name: Setup go deps
|
||||||
|
run: |
|
||||||
|
# Install golangci-lint
|
||||||
|
curl -sSfL https://raw.githubusercontent.com/golangci/golangci-lint/HEAD/install.sh | sh -s -- -b $(go env GOPATH)/bin
|
||||||
|
|
||||||
|
# Install go-task dependencies
|
||||||
|
go install golang.org/x/tools/cmd/goimports@latest
|
||||||
|
|
||||||
|
- name: Run goimports
|
||||||
|
run: goimports -w .
|
||||||
|
|
||||||
|
- name: Run golangci-lint autofix
|
||||||
|
run: golangci-lint run --fix
|
||||||
|
|
||||||
|
- name: Run golangci-lint format
|
||||||
|
run: golangci-lint fmt
|
||||||
|
|
||||||
|
- name: Run go mod tidy
|
||||||
|
run: go mod tidy
|
||||||
|
|
||||||
|
- name: Run gopls modernize
|
||||||
|
run: task modernize
|
||||||
|
|
||||||
|
- uses: autofix-ci/action@v1
|
||||||
|
|||||||
508
.github/workflows/ci.yml
vendored
508
.github/workflows/ci.yml
vendored
@ -2,9 +2,8 @@ name: CI
|
|||||||
|
|
||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
branches: ['master', 'develop']
|
branches: ["master", "develop"]
|
||||||
pull_request:
|
pull_request:
|
||||||
branches: ['master', 'develop', 'feature/*']
|
|
||||||
|
|
||||||
env:
|
env:
|
||||||
REGISTRY: ghcr.io
|
REGISTRY: ghcr.io
|
||||||
@ -15,81 +14,71 @@ concurrency:
|
|||||||
cancel-in-progress: true
|
cancel-in-progress: true
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
|
golangci:
|
||||||
|
name: lint
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
pull-requests: read
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v6
|
||||||
|
- uses: actions/setup-go@v6
|
||||||
|
with:
|
||||||
|
go-version: stable
|
||||||
|
- name: golangci-lint
|
||||||
|
uses: golangci/golangci-lint-action@v9
|
||||||
|
with: { version: latest }
|
||||||
|
|
||||||
test:
|
test:
|
||||||
name: Test
|
name: Test
|
||||||
|
needs: [golangci]
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
permissions:
|
permissions:
|
||||||
contents: write
|
contents: write
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
go:
|
go:
|
||||||
- 1.21.x
|
|
||||||
- 1.22.x
|
|
||||||
- 1.23.x
|
|
||||||
- 1.24.x
|
- 1.24.x
|
||||||
|
- 1.25.x
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v6
|
||||||
|
|
||||||
- name: Set up Go ${{ matrix.go }}
|
- name: Set up Go ${{ matrix.go }}
|
||||||
uses: actions/setup-go@v5
|
uses: actions/setup-go@v6
|
||||||
with:
|
with:
|
||||||
go-version: ${{ matrix.go }}
|
go-version: ${{ matrix.go }}
|
||||||
check-latest: true
|
check-latest: true
|
||||||
|
|
||||||
- name: Download dependencies with retry
|
- name: Install Task
|
||||||
run: |
|
uses: go-task/setup-task@v1
|
||||||
set -e
|
|
||||||
echo "Downloading Go dependencies..."
|
|
||||||
|
|
||||||
# Function to download with retry
|
- name: Show build info
|
||||||
download_with_retry() {
|
run: task info
|
||||||
local attempt=1
|
|
||||||
local max_attempts=3
|
|
||||||
|
|
||||||
while [ $attempt -le $max_attempts ]; do
|
- name: Download dependencies
|
||||||
echo "Attempt $attempt of $max_attempts"
|
run: task deps
|
||||||
|
|
||||||
if go mod download; then
|
|
||||||
echo "Download successful on attempt $attempt"
|
|
||||||
return 0
|
|
||||||
else
|
|
||||||
echo "Download failed on attempt $attempt"
|
|
||||||
if [ $attempt -lt $max_attempts ]; then
|
|
||||||
echo "Cleaning cache and retrying..."
|
|
||||||
go clean -modcache
|
|
||||||
go clean -cache
|
|
||||||
sleep 2
|
|
||||||
fi
|
|
||||||
attempt=$((attempt + 1))
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
|
|
||||||
echo "All download attempts failed"
|
|
||||||
return 1
|
|
||||||
}
|
|
||||||
|
|
||||||
# Try download with retry logic
|
|
||||||
download_with_retry
|
|
||||||
|
|
||||||
echo "Verifying module dependencies..."
|
|
||||||
go mod verify
|
|
||||||
echo "Dependencies verified successfully"
|
|
||||||
|
|
||||||
- name: Build
|
- name: Build
|
||||||
run: go build -v ./...
|
run: task build
|
||||||
|
|
||||||
- name: Run tests with enhanced reporting
|
- name: Run tests with enhanced reporting
|
||||||
id: test
|
id: test
|
||||||
|
env:
|
||||||
|
CGO_ENABLED: 1
|
||||||
run: |
|
run: |
|
||||||
echo "## 🔧 Test Environment" >> $GITHUB_STEP_SUMMARY
|
{
|
||||||
echo "- **Go Version:** ${{ matrix.go }}" >> $GITHUB_STEP_SUMMARY
|
cat << EOF
|
||||||
echo "- **OS:** ubuntu-latest" >> $GITHUB_STEP_SUMMARY
|
## 🔧 Test Environment
|
||||||
echo "- **Timestamp:** $(date -u)" >> $GITHUB_STEP_SUMMARY
|
- **Go Version:** ${{ matrix.go }}
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
- **OS:** ubuntu-latest
|
||||||
|
- **Timestamp:** $(date -u)
|
||||||
|
|
||||||
|
EOF
|
||||||
|
} >> "$GITHUB_STEP_SUMMARY"
|
||||||
|
|
||||||
echo "Running tests with coverage..."
|
echo "Running tests with coverage..."
|
||||||
go test -v -race -coverprofile=coverage.out ./... 2>&1 | tee test-output.log
|
task test:coverage 2>&1 | tee test-output.log
|
||||||
|
|
||||||
# Extract test results for summary
|
# Extract test results for summary
|
||||||
TEST_STATUS=$?
|
TEST_STATUS=$?
|
||||||
@ -99,183 +88,202 @@ jobs:
|
|||||||
SKIPPED_TESTS=$(grep -c "--- SKIP:" test-output.log || echo "0")
|
SKIPPED_TESTS=$(grep -c "--- SKIP:" test-output.log || echo "0")
|
||||||
|
|
||||||
# Generate test summary
|
# Generate test summary
|
||||||
echo "## 🧪 Test Results (Go ${{ matrix.go }})" >> $GITHUB_STEP_SUMMARY
|
{
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
cat << EOF
|
||||||
echo "| Metric | Value |" >> $GITHUB_STEP_SUMMARY
|
## 🧪 Test Results (Go ${{ matrix.go }})
|
||||||
echo "|--------|-------|" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "| Total Tests | $TOTAL_TESTS |" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "| Passed | ✅ $PASSED_TESTS |" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "| Failed | ❌ $FAILED_TESTS |" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "| Skipped | ⏭️ $SKIPPED_TESTS |" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "| Status | $([ $TEST_STATUS -eq 0 ] && echo "✅ PASSED" || echo "❌ FAILED") |" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
|
|
||||||
# Add package breakdown
|
| Metric | Value |
|
||||||
echo "### 📦 Package Test Results" >> $GITHUB_STEP_SUMMARY
|
| ----------- | ------------------------------------------------------------- |
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
| Total Tests | $TOTAL_TESTS |
|
||||||
echo "| Package | Status |" >> $GITHUB_STEP_SUMMARY
|
| Passed | $PASSED_TESTS |
|
||||||
echo "|---------|--------|" >> $GITHUB_STEP_SUMMARY
|
| Failed | $FAILED_TESTS |
|
||||||
|
| Skipped | $SKIPPED_TESTS |
|
||||||
|
| Status | $([ "$TEST_STATUS" -eq 0 ] && echo "PASSED" || echo "FAILED") |
|
||||||
|
|
||||||
# Extract package results
|
### 📦 Package Test Results
|
||||||
grep "^ok\|^FAIL" test-output.log | while read line; do
|
|
||||||
if [[ $line == ok* ]]; then
|
| Package | Status |
|
||||||
pkg=$(echo $line | awk '{print $2}')
|
|---------|--------|
|
||||||
echo "| $pkg | ✅ PASS |" >> $GITHUB_STEP_SUMMARY
|
EOF
|
||||||
elif [[ $line == FAIL* ]]; then
|
|
||||||
pkg=$(echo $line | awk '{print $2}')
|
# Extract package results
|
||||||
echo "| $pkg | ❌ FAIL |" >> $GITHUB_STEP_SUMMARY
|
grep "^ok\|^FAIL" test-output.log | while read -r line; do
|
||||||
|
if [[ $line == ok* ]]; then
|
||||||
|
pkg=$(echo "$line" | awk '{print $2}')
|
||||||
|
echo "| $pkg | ✅ PASS |"
|
||||||
|
elif [[ $line == FAIL* ]]; then
|
||||||
|
pkg=$(echo "$line" | awk '{print $2}')
|
||||||
|
echo "| $pkg | ❌ FAIL |"
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Add detailed results if tests failed
|
||||||
|
if [ "$TEST_STATUS" -ne 0 ]; then
|
||||||
|
cat << 'EOF'
|
||||||
|
### ❌ Failed Tests Details
|
||||||
|
```
|
||||||
|
EOF
|
||||||
|
grep -A 10 -- "--- FAIL:" test-output.log | head -100
|
||||||
|
cat << 'EOF'
|
||||||
|
```
|
||||||
|
|
||||||
|
EOF
|
||||||
fi
|
fi
|
||||||
done
|
} >> "$GITHUB_STEP_SUMMARY"
|
||||||
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
|
|
||||||
# Add detailed results if tests failed
|
|
||||||
if [ $TEST_STATUS -ne 0 ]; then
|
|
||||||
echo "### ❌ Failed Tests Details" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
grep -A 10 "--- FAIL:" test-output.log | head -100 >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Set outputs for other steps
|
# Set outputs for other steps
|
||||||
echo "test-status=$TEST_STATUS" >> $GITHUB_OUTPUT
|
{
|
||||||
echo "total-tests=$TOTAL_TESTS" >> $GITHUB_OUTPUT
|
echo "test-status=$TEST_STATUS"
|
||||||
echo "passed-tests=$PASSED_TESTS" >> $GITHUB_OUTPUT
|
echo "total-tests=$TOTAL_TESTS"
|
||||||
echo "failed-tests=$FAILED_TESTS" >> $GITHUB_OUTPUT
|
echo "passed-tests=$PASSED_TESTS"
|
||||||
|
echo "failed-tests=$FAILED_TESTS"
|
||||||
|
} >> "$GITHUB_OUTPUT"
|
||||||
|
|
||||||
# Exit with the original test status
|
# Exit with the original test status
|
||||||
exit $TEST_STATUS
|
exit "$TEST_STATUS"
|
||||||
|
|
||||||
- name: Generate coverage report
|
- name: Generate coverage report
|
||||||
if: always()
|
if: always()
|
||||||
run: |
|
run: |
|
||||||
if [ -f coverage.out ]; then
|
if [ -f coverage/coverage.out ]; then
|
||||||
go tool cover -html=coverage.out -o coverage.html
|
COVERAGE=$(go tool cover -func=coverage/coverage.out | grep total | awk '{print $3}')
|
||||||
COVERAGE=$(go tool cover -func=coverage.out | grep total | awk '{print $3}')
|
|
||||||
|
|
||||||
echo "## 📊 Code Coverage (Go ${{ matrix.go }})" >> $GITHUB_STEP_SUMMARY
|
{
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
cat << EOF
|
||||||
echo "**Total Coverage: $COVERAGE**" >> $GITHUB_STEP_SUMMARY
|
## 📊 Code Coverage (Go ${{ matrix.go }})
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
|
|
||||||
# Add coverage by package
|
**Total Coverage: $COVERAGE**
|
||||||
echo "<details>" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "<summary>Click to expand 📋 Coverage by Package details</summary>" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "| Package | Coverage |" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "|---------|----------|" >> $GITHUB_STEP_SUMMARY
|
|
||||||
|
|
||||||
# Create temporary file for package coverage aggregation
|
<details>
|
||||||
temp_coverage=$(mktemp)
|
<summary>Click to expand 📋 Coverage by Package details</summary>
|
||||||
|
|
||||||
# Extract package-level coverage data
|
| Package | Coverage |
|
||||||
go tool cover -func=coverage.out | grep -v total | while read line; do
|
| ------- | -------- |
|
||||||
if [[ $line == *".go:"* ]]; then
|
EOF
|
||||||
# Extract package path from file path (everything before the filename)
|
|
||||||
filepath=$(echo "$line" | awk '{print $1}')
|
|
||||||
pkg_path=$(dirname "$filepath" | sed 's|github.com/kjanat/articulate-parser/||' | sed 's|^\./||')
|
|
||||||
coverage=$(echo "$line" | awk '{print $3}' | sed 's/%//')
|
|
||||||
|
|
||||||
# Use root package if no subdirectory
|
# Create temporary file for package coverage aggregation
|
||||||
if [[ "$pkg_path" == "." || "$pkg_path" == "" ]]; then
|
temp_coverage=$(mktemp)
|
||||||
pkg_path="root"
|
|
||||||
|
# Extract package-level coverage data
|
||||||
|
go tool cover -func=coverage/coverage.out | grep -v total | while read -r line; do
|
||||||
|
if [[ $line == *".go:"* ]]; then
|
||||||
|
# Extract package path from file path (everything before the filename)
|
||||||
|
filepath=$(echo "$line" | awk '{print $1}')
|
||||||
|
pkg_path=$(dirname "$filepath" | sed 's|github.com/kjanat/articulate-parser/||; s|^\./||')
|
||||||
|
coverage=$(echo "$line" | awk '{print $3}' | sed 's/%//')
|
||||||
|
|
||||||
|
# Use root package if no subdirectory
|
||||||
|
[[ "$pkg_path" == "." || "$pkg_path" == "" ]] && pkg_path="root"
|
||||||
|
|
||||||
|
echo "$pkg_path $coverage" >> "$temp_coverage"
|
||||||
fi
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
echo "$pkg_path $coverage" >> "$temp_coverage"
|
# Aggregate coverage by package (average)
|
||||||
fi
|
awk '{
|
||||||
done
|
packages[$1] += $2
|
||||||
|
counts[$1]++
|
||||||
# Aggregate coverage by package (average)
|
|
||||||
awk '{
|
|
||||||
packages[$1] += $2;
|
|
||||||
counts[$1]++
|
|
||||||
}
|
|
||||||
END {
|
|
||||||
for (pkg in packages) {
|
|
||||||
avg = packages[pkg] / counts[pkg]
|
|
||||||
printf "| %s | %.1f%% |\n", pkg, avg
|
|
||||||
}
|
}
|
||||||
}' $temp_coverage | sort >> $GITHUB_STEP_SUMMARY
|
END {
|
||||||
|
for (pkg in packages) {
|
||||||
|
avg = packages[pkg] / counts[pkg]
|
||||||
|
printf "| %s | %.1f%% |\n", pkg, avg
|
||||||
|
}
|
||||||
|
}' "$temp_coverage" | sort
|
||||||
|
|
||||||
rm -f $temp_coverage
|
rm -f "$temp_coverage"
|
||||||
|
|
||||||
echo "</details>" >> $GITHUB_STEP_SUMMARY
|
cat << 'EOF'
|
||||||
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
</details>
|
||||||
|
|
||||||
|
EOF
|
||||||
|
} >> "$GITHUB_STEP_SUMMARY"
|
||||||
else
|
else
|
||||||
echo "## ⚠️ Coverage Report" >> $GITHUB_STEP_SUMMARY
|
cat >> "$GITHUB_STEP_SUMMARY" << 'EOF'
|
||||||
echo "No coverage file generated" >> $GITHUB_STEP_SUMMARY
|
## ⚠️ Coverage Report
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
No coverage file generated
|
||||||
|
|
||||||
|
EOF
|
||||||
fi
|
fi
|
||||||
|
|
||||||
- name: Upload test artifacts
|
- name: Upload test artifacts
|
||||||
if: failure()
|
if: failure()
|
||||||
uses: actions/upload-artifact@v4
|
uses: actions/upload-artifact@v6
|
||||||
with:
|
with:
|
||||||
name: test-results-go-${{ matrix.go }}
|
name: test-results-go-${{ matrix.go }}
|
||||||
path: |
|
path: |
|
||||||
test-output.log
|
test-output.log
|
||||||
coverage.out
|
coverage/
|
||||||
coverage.html
|
|
||||||
retention-days: 7
|
retention-days: 7
|
||||||
|
|
||||||
- name: Run go vet
|
- name: Run linters
|
||||||
run: |
|
run: |
|
||||||
echo "## 🔍 Static Analysis (Go ${{ matrix.go }})" >> $GITHUB_STEP_SUMMARY
|
{
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
cat << EOF
|
||||||
|
## 🔍 Static Analysis (Go ${{ matrix.go }})
|
||||||
|
|
||||||
VET_OUTPUT=$(go vet ./... 2>&1 || echo "")
|
EOF
|
||||||
VET_STATUS=$?
|
|
||||||
|
|
||||||
if [ $VET_STATUS -eq 0 ]; then
|
# Run go vet
|
||||||
echo "✅ **go vet:** No issues found" >> $GITHUB_STEP_SUMMARY
|
VET_OUTPUT=$(task lint:vet 2>&1 || echo "")
|
||||||
else
|
VET_STATUS=$?
|
||||||
echo "❌ **go vet:** Issues found" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "$VET_OUTPUT" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
fi
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
|
|
||||||
exit $VET_STATUS
|
if [ "$VET_STATUS" -eq 0 ]; then
|
||||||
|
echo "✅ **go vet:** No issues found"
|
||||||
|
else
|
||||||
|
cat << 'EOF'
|
||||||
|
❌ **go vet:** Issues found
|
||||||
|
|
||||||
- name: Run go fmt
|
```
|
||||||
run: |
|
EOF
|
||||||
FMT_OUTPUT=$(gofmt -s -l . 2>&1 || echo "")
|
echo "$VET_OUTPUT"
|
||||||
|
echo '```'
|
||||||
|
fi
|
||||||
|
echo ""
|
||||||
|
|
||||||
if [ -z "$FMT_OUTPUT" ]; then
|
# Run go fmt check
|
||||||
echo "✅ **go fmt:** All files properly formatted" >> $GITHUB_STEP_SUMMARY
|
FMT_OUTPUT=$(task lint:fmt 2>&1 || echo "")
|
||||||
else
|
FMT_STATUS=$?
|
||||||
echo "❌ **go fmt:** Files need formatting" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
if [ "$FMT_STATUS" -eq 0 ]; then
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
echo "✅ **go fmt:** All files properly formatted"
|
||||||
echo "$FMT_OUTPUT" >> $GITHUB_STEP_SUMMARY
|
else
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
cat << 'EOF'
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
❌ **go fmt:** Files need formatting
|
||||||
exit 1
|
|
||||||
fi
|
```
|
||||||
|
EOF
|
||||||
|
echo "$FMT_OUTPUT"
|
||||||
|
echo '```'
|
||||||
|
fi
|
||||||
|
} >> "$GITHUB_STEP_SUMMARY"
|
||||||
|
|
||||||
|
# Exit with error if any linter failed
|
||||||
|
[ "$VET_STATUS" -eq 0 ] && [ "$FMT_STATUS" -eq 0 ] || exit 1
|
||||||
|
|
||||||
- name: Job Summary
|
- name: Job Summary
|
||||||
if: always()
|
if: always()
|
||||||
run: |
|
run: |
|
||||||
echo "## 📋 Job Summary (Go ${{ matrix.go }})" >> $GITHUB_STEP_SUMMARY
|
cat >> "$GITHUB_STEP_SUMMARY" << 'EOF'
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
## 📋 Job Summary (Go ${{ matrix.go }})
|
||||||
echo "| Step | Status |" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "|------|--------|" >> $GITHUB_STEP_SUMMARY
|
| Step | Status |
|
||||||
echo "| Dependencies | ✅ Success |" >> $GITHUB_STEP_SUMMARY
|
| --------------- | --------------------------------------------------------------- |
|
||||||
echo "| Build | ✅ Success |" >> $GITHUB_STEP_SUMMARY
|
| Dependencies | Success |
|
||||||
echo "| Tests | ${{ steps.test.outcome == 'success' && '✅ Success' || '❌ Failed' }} |" >> $GITHUB_STEP_SUMMARY
|
| Build | Success |
|
||||||
echo "| Coverage | ${{ job.status == 'success' && '✅ Generated' || '⚠️ Partial' }} |" >> $GITHUB_STEP_SUMMARY
|
| Tests | ${{ steps.test.outcome == 'success' && 'Success' || 'Failed' }} |
|
||||||
echo "| Static Analysis | ${{ job.status == 'success' && '✅ Clean' || '❌ Issues' }} |" >> $GITHUB_STEP_SUMMARY
|
| Coverage | ${{ job.status == 'success' && 'Generated' || 'Partial' }} |
|
||||||
echo "| Code Formatting | ${{ job.status == 'success' && '✅ Clean' || '❌ Issues' }} |" >> $GITHUB_STEP_SUMMARY
|
| Static Analysis | ${{ job.status == 'success' && 'Clean' || 'Issues' }} |
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
| Code Formatting | ${{ job.status == 'success' && 'Clean' || 'Issues' }} |
|
||||||
|
EOF
|
||||||
|
|
||||||
- name: Upload coverage reports to Codecov
|
- name: Upload coverage reports to Codecov
|
||||||
uses: codecov/codecov-action@v5
|
uses: codecov/codecov-action@v5
|
||||||
with:
|
with:
|
||||||
|
files: ./coverage/coverage.out
|
||||||
flags: Go ${{ matrix.go }}
|
flags: Go ${{ matrix.go }}
|
||||||
slug: kjanat/articulate-parser
|
slug: kjanat/articulate-parser
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
@ -295,44 +303,45 @@ jobs:
|
|||||||
contents: read
|
contents: read
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v6
|
||||||
|
|
||||||
|
- name: Set up Go
|
||||||
|
uses: actions/setup-go@v6
|
||||||
|
with:
|
||||||
|
go-version-file: go.mod
|
||||||
|
check-latest: true
|
||||||
|
|
||||||
|
- name: Install Task
|
||||||
|
uses: go-task/setup-task@v1
|
||||||
|
|
||||||
- name: Set up Docker Buildx
|
- name: Set up Docker Buildx
|
||||||
uses: docker/setup-buildx-action@v3
|
uses: docker/setup-buildx-action@v3
|
||||||
|
|
||||||
- name: Capture build date
|
- name: Build Docker image using Task
|
||||||
run: echo "BUILD_TIME=$(git log -1 --format=%cd --date=iso-strict)" >> $GITHUB_ENV
|
run: task docker:build
|
||||||
|
|
||||||
- name: Build Docker image (test)
|
- name: Test Docker image using Task
|
||||||
uses: docker/build-push-action@v6
|
|
||||||
with:
|
|
||||||
context: .
|
|
||||||
push: false
|
|
||||||
load: true
|
|
||||||
tags: test:latest
|
|
||||||
build-args: |
|
|
||||||
VERSION=test
|
|
||||||
BUILD_TIME=${{ env.BUILD_TIME }}
|
|
||||||
GIT_COMMIT=${{ github.sha }}
|
|
||||||
cache-from: type=gha
|
|
||||||
cache-to: type=gha,mode=max
|
|
||||||
|
|
||||||
- name: Test Docker image
|
|
||||||
run: |
|
run: |
|
||||||
echo "## 🧪 Docker Image Tests" >> $GITHUB_STEP_SUMMARY
|
{
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
cat << 'EOF'
|
||||||
|
## 🧪 Docker Image Tests
|
||||||
|
|
||||||
# Test that the image runs and shows help
|
EOF
|
||||||
echo "**Testing help command:**" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
docker run --rm test:latest --help >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
|
|
||||||
# Test image size
|
# Run Task docker test
|
||||||
IMAGE_SIZE=$(docker image inspect test:latest --format='{{.Size}}' | numfmt --to=iec-i --suffix=B)
|
task docker:test
|
||||||
echo "**Image size:** $IMAGE_SIZE" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
echo "**Testing help command:**"
|
||||||
|
echo '```terminaloutput'
|
||||||
|
docker run --rm articulate-parser:latest --help
|
||||||
|
echo '```'
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Test image size
|
||||||
|
IMAGE_SIZE=$(docker image inspect articulate-parser:latest --format='{{.Size}}' | numfmt --to=iec-i --suffix=B)
|
||||||
|
echo "**Image size:** $IMAGE_SIZE"
|
||||||
|
echo ""
|
||||||
|
} >> "$GITHUB_STEP_SUMMARY"
|
||||||
|
|
||||||
dependency-review:
|
dependency-review:
|
||||||
name: Dependency Review
|
name: Dependency Review
|
||||||
@ -341,28 +350,29 @@ jobs:
|
|||||||
contents: read
|
contents: read
|
||||||
if: github.event_name == 'pull_request'
|
if: github.event_name == 'pull_request'
|
||||||
steps:
|
steps:
|
||||||
- name: 'Checkout Repository'
|
- name: "Checkout Repository"
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v6
|
||||||
|
|
||||||
- name: 'Dependency Review'
|
- name: "Dependency Review"
|
||||||
uses: actions/dependency-review-action@v4
|
uses: actions/dependency-review-action@v4
|
||||||
with:
|
with:
|
||||||
fail-on-severity: moderate
|
fail-on-severity: moderate
|
||||||
comment-summary-in-pr: always
|
comment-summary-in-pr: always
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
docker:
|
docker:
|
||||||
name: Docker Build & Push
|
name: Docker Build & Push
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
permissions:
|
permissions:
|
||||||
contents: read
|
contents: read
|
||||||
packages: write
|
packages: write
|
||||||
needs: ['test']
|
needs: [test, docker-test]
|
||||||
if: github.event_name == 'push' && (github.ref == 'refs/heads/master' || github.ref == 'refs/heads/develop' || startsWith(github.ref, 'refs/heads/feature/docker'))
|
if: |
|
||||||
|
github.event_name == 'push' && (github.ref == 'refs/heads/master' ||
|
||||||
|
github.ref == 'refs/heads/develop' ||
|
||||||
|
startsWith(github.ref, 'refs/heads/feature/docker'))
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v6
|
||||||
|
|
||||||
- name: Login to Docker Hub
|
- name: Login to Docker Hub
|
||||||
uses: docker/login-action@v3
|
uses: docker/login-action@v3
|
||||||
@ -434,35 +444,39 @@ jobs:
|
|||||||
|
|
||||||
- name: Generate Docker summary
|
- name: Generate Docker summary
|
||||||
run: |
|
run: |
|
||||||
echo "## 🐳 Docker Build Summary" >> $GITHUB_STEP_SUMMARY
|
cat >> "$GITHUB_STEP_SUMMARY" << 'EOF'
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
## 🐳 Docker Build Summary
|
||||||
echo "**Image:** \`ghcr.io/${{ github.repository }}\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
**Image:** `ghcr.io/${{ github.repository }}`
|
||||||
echo "**Tags built:**" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
**Tags built:**
|
||||||
echo "${{ steps.meta.outputs.tags }}" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
```text
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
${{ steps.meta.outputs.tags }}
|
||||||
echo "**Features:**" >> $GITHUB_STEP_SUMMARY
|
```
|
||||||
echo "- **Platforms:** linux/amd64, linux/arm64, linux/arm/v7, linux/386, linux/ppc64le, linux/s390x" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "- **Architecture optimization:** ✅ Native compilation for each platform" >> $GITHUB_STEP_SUMMARY
|
**Features:**
|
||||||
echo "- **Multi-arch image description:** ✅ Enabled" >> $GITHUB_STEP_SUMMARY
|
- **Platforms:** linux/amd64, linux/arm64, linux/arm/v7, linux/386, linux/ppc64le, linux/s390x
|
||||||
echo "- **SBOM (Software Bill of Materials):** ✅ Generated" >> $GITHUB_STEP_SUMMARY
|
- **Architecture optimization:** Native compilation for each platform
|
||||||
echo "- **Provenance attestation:** ✅ Generated" >> $GITHUB_STEP_SUMMARY
|
- **Multi-arch image description:** Enabled
|
||||||
echo "- **Security scanning:** Ready for vulnerability analysis" >> $GITHUB_STEP_SUMMARY
|
- **SBOM (Software Bill of Materials):** Generated
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
- **Provenance attestation:** Generated
|
||||||
echo "**Usage:**" >> $GITHUB_STEP_SUMMARY
|
- **Security scanning:** Ready for vulnerability analysis
|
||||||
echo "\`\`\`bash" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "# Pull the image" >> $GITHUB_STEP_SUMMARY
|
**Usage:**
|
||||||
echo "docker pull ghcr.io/${{ github.repository }}:latest" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
```bash
|
||||||
echo "# Run with help" >> $GITHUB_STEP_SUMMARY
|
# Pull the image
|
||||||
echo "docker run --rm ghcr.io/${{ github.repository }}:latest --help" >> $GITHUB_STEP_SUMMARY
|
docker pull ghcr.io/${{ github.repository }}:latest
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "# Process a local file (mount current directory)" >> $GITHUB_STEP_SUMMARY
|
# Run with help
|
||||||
echo "docker run --rm -v \$(pwd):/workspace ghcr.io/${{ github.repository }}:latest /workspace/input.json markdown /workspace/output.md" >> $GITHUB_STEP_SUMMARY
|
docker run --rm ghcr.io/${{ github.repository }}:latest --help
|
||||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "" >> $GITHUB_STEP_SUMMARY
|
# Process a local file (mount current directory)
|
||||||
|
docker run --rm -v $(pwd):/workspace ghcr.io/${{ github.repository }}:latest /workspace/input.json markdown /workspace/output.md
|
||||||
|
```
|
||||||
|
|
||||||
|
EOF
|
||||||
|
|
||||||
# Security and quality analysis workflows
|
# Security and quality analysis workflows
|
||||||
codeql-analysis:
|
codeql-analysis:
|
||||||
|
|||||||
100
.github/workflows/codeql.yml
vendored
100
.github/workflows/codeql.yml
vendored
@ -17,7 +17,7 @@ name: "CodeQL"
|
|||||||
on:
|
on:
|
||||||
workflow_call:
|
workflow_call:
|
||||||
schedule:
|
schedule:
|
||||||
- cron: '44 16 * * 6'
|
- cron: "44 16 * * 6"
|
||||||
# push:
|
# push:
|
||||||
# branches: [ "master" ]
|
# branches: [ "master" ]
|
||||||
# pull_request:
|
# pull_request:
|
||||||
@ -47,58 +47,58 @@ jobs:
|
|||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
include:
|
include:
|
||||||
- language: actions
|
- language: actions
|
||||||
build-mode: none
|
build-mode: none
|
||||||
- language: go
|
- language: go
|
||||||
build-mode: autobuild
|
build-mode: autobuild
|
||||||
# CodeQL supports the following values keywords for 'language': 'actions', 'c-cpp', 'csharp', 'go', 'java-kotlin', 'javascript-typescript', 'python', 'ruby', 'swift'
|
# CodeQL supports the following values keywords for 'language': 'actions', 'c-cpp', 'csharp', 'go', 'java-kotlin', 'javascript-typescript', 'python', 'ruby', 'swift'
|
||||||
# Use `c-cpp` to analyze code written in C, C++ or both
|
# Use `c-cpp` to analyze code written in C, C++ or both
|
||||||
# Use 'java-kotlin' to analyze code written in Java, Kotlin or both
|
# Use 'java-kotlin' to analyze code written in Java, Kotlin or both
|
||||||
# Use 'javascript-typescript' to analyze code written in JavaScript, TypeScript or both
|
# Use 'javascript-typescript' to analyze code written in JavaScript, TypeScript or both
|
||||||
# To learn more about changing the languages that are analyzed or customizing the build mode for your analysis,
|
# To learn more about changing the languages that are analyzed or customizing the build mode for your analysis,
|
||||||
# see https://docs.github.com/en/code-security/code-scanning/creating-an-advanced-setup-for-code-scanning/customizing-your-advanced-setup-for-code-scanning.
|
# see https://docs.github.com/en/code-security/code-scanning/creating-an-advanced-setup-for-code-scanning/customizing-your-advanced-setup-for-code-scanning.
|
||||||
# If you are analyzing a compiled language, you can modify the 'build-mode' for that language to customize how
|
# If you are analyzing a compiled language, you can modify the 'build-mode' for that language to customize how
|
||||||
# your codebase is analyzed, see https://docs.github.com/en/code-security/code-scanning/creating-an-advanced-setup-for-code-scanning/codeql-code-scanning-for-compiled-languages
|
# your codebase is analyzed, see https://docs.github.com/en/code-security/code-scanning/creating-an-advanced-setup-for-code-scanning/codeql-code-scanning-for-compiled-languages
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v6
|
||||||
|
|
||||||
# Add any setup steps before running the `github/codeql-action/init` action.
|
# Add any setup steps before running the `github/codeql-action/init` action.
|
||||||
# This includes steps like installing compilers or runtimes (`actions/setup-node`
|
# This includes steps like installing compilers or runtimes (`actions/setup-node`
|
||||||
# or others). This is typically only required for manual builds.
|
# or others). This is typically only required for manual builds.
|
||||||
# - name: Setup runtime (example)
|
# - name: Setup runtime (example)
|
||||||
# uses: actions/setup-example@v1
|
# uses: actions/setup-example@v1
|
||||||
|
|
||||||
# Initializes the CodeQL tools for scanning.
|
# Initializes the CodeQL tools for scanning.
|
||||||
- name: Initialize CodeQL
|
- name: Initialize CodeQL
|
||||||
uses: github/codeql-action/init@v3
|
uses: github/codeql-action/init@v4
|
||||||
with:
|
with:
|
||||||
languages: ${{ matrix.language }}
|
languages: ${{ matrix.language }}
|
||||||
build-mode: ${{ matrix.build-mode }}
|
build-mode: ${{ matrix.build-mode }}
|
||||||
# If you wish to specify custom queries, you can do so here or in a config file.
|
# If you wish to specify custom queries, you can do so here or in a config file.
|
||||||
# By default, queries listed here will override any specified in a config file.
|
# By default, queries listed here will override any specified in a config file.
|
||||||
# Prefix the list here with "+" to use these queries and those in the config file.
|
# Prefix the list here with "+" to use these queries and those in the config file.
|
||||||
|
|
||||||
# For more details on CodeQL's query packs, refer to: https://docs.github.com/en/code-security/code-scanning/automatically-scanning-your-code-for-vulnerabilities-and-errors/configuring-code-scanning#using-queries-in-ql-packs
|
# For more details on CodeQL's query packs, refer to: https://docs.github.com/en/code-security/code-scanning/automatically-scanning-your-code-for-vulnerabilities-and-errors/configuring-code-scanning#using-queries-in-ql-packs
|
||||||
# queries: security-extended,security-and-quality
|
# queries: security-extended,security-and-quality
|
||||||
|
|
||||||
# If the analyze step fails for one of the languages you are analyzing with
|
# If the analyze step fails for one of the languages you are analyzing with
|
||||||
# "We were unable to automatically build your code", modify the matrix above
|
# "We were unable to automatically build your code", modify the matrix above
|
||||||
# to set the build mode to "manual" for that language. Then modify this step
|
# to set the build mode to "manual" for that language. Then modify this step
|
||||||
# to build your code.
|
# to build your code.
|
||||||
# ℹ️ Command-line programs to run using the OS shell.
|
# ℹ️ Command-line programs to run using the OS shell.
|
||||||
# 📚 See https://docs.github.com/en/actions/using-workflows/workflow-syntax-for-github-actions#jobsjob_idstepsrun
|
# 📚 See https://docs.github.com/en/actions/using-workflows/workflow-syntax-for-github-actions#jobsjob_idstepsrun
|
||||||
- if: matrix.build-mode == 'manual'
|
- if: matrix.build-mode == 'manual'
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
echo 'If you are using a "manual" build mode for one or more of the' \
|
echo 'If you are using a "manual" build mode for one or more of the' \
|
||||||
'languages you are analyzing, replace this with the commands to build' \
|
'languages you are analyzing, replace this with the commands to build' \
|
||||||
'your code, for example:'
|
'your code, for example:'
|
||||||
echo ' make bootstrap'
|
echo ' make bootstrap'
|
||||||
echo ' make release'
|
echo ' make release'
|
||||||
exit 1
|
exit 1
|
||||||
|
|
||||||
- name: Perform CodeQL Analysis
|
- name: Perform CodeQL Analysis
|
||||||
uses: github/codeql-action/analyze@v3
|
uses: github/codeql-action/analyze@v4
|
||||||
with:
|
with:
|
||||||
category: "/language:${{matrix.language}}"
|
category: "/language:${{matrix.language}}"
|
||||||
|
|||||||
6
.github/workflows/dependency-review.yml
vendored
6
.github/workflows/dependency-review.yml
vendored
@ -16,10 +16,10 @@ jobs:
|
|||||||
dependency-review:
|
dependency-review:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- name: 'Checkout Repository'
|
- name: "Checkout Repository"
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v6
|
||||||
|
|
||||||
- name: 'Dependency Review'
|
- name: "Dependency Review"
|
||||||
uses: actions/dependency-review-action@v4
|
uses: actions/dependency-review-action@v4
|
||||||
with:
|
with:
|
||||||
fail-on-severity: moderate
|
fail-on-severity: moderate
|
||||||
|
|||||||
12
.github/workflows/release.yml
vendored
12
.github/workflows/release.yml
vendored
@ -3,7 +3,7 @@ name: Release
|
|||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
tags:
|
tags:
|
||||||
- 'v*.*.*'
|
- "v*.*.*"
|
||||||
workflow_call:
|
workflow_call:
|
||||||
|
|
||||||
env:
|
env:
|
||||||
@ -20,14 +20,14 @@ jobs:
|
|||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout code
|
- name: Checkout code
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v6
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
|
|
||||||
- name: Set up Go
|
- name: Set up Go
|
||||||
uses: actions/setup-go@v5
|
uses: actions/setup-go@v6
|
||||||
with:
|
with:
|
||||||
go-version-file: 'go.mod'
|
go-version-file: "go.mod"
|
||||||
check-latest: true
|
check-latest: true
|
||||||
|
|
||||||
- name: Run tests
|
- name: Run tests
|
||||||
@ -82,13 +82,13 @@ jobs:
|
|||||||
docker:
|
docker:
|
||||||
name: Docker Build & Push
|
name: Docker Build & Push
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
needs: ['release']
|
needs: ["release"]
|
||||||
permissions:
|
permissions:
|
||||||
contents: read
|
contents: read
|
||||||
packages: write
|
packages: write
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v6
|
||||||
|
|
||||||
- name: Login to Docker Hub
|
- name: Login to Docker Hub
|
||||||
uses: docker/login-action@v3
|
uses: docker/login-action@v3
|
||||||
|
|||||||
8
.gitignore
vendored
8
.gitignore
vendored
@ -69,3 +69,11 @@ main_coverage
|
|||||||
# Editors
|
# Editors
|
||||||
.vscode/
|
.vscode/
|
||||||
.idea/
|
.idea/
|
||||||
|
|
||||||
|
.task/
|
||||||
|
|
||||||
|
**/*.local.*
|
||||||
|
|
||||||
|
.claude/
|
||||||
|
|
||||||
|
NUL
|
||||||
|
|||||||
396
.golangci.yml
Normal file
396
.golangci.yml
Normal file
@ -0,0 +1,396 @@
|
|||||||
|
# golangci-lint configuration for articulate-parser
|
||||||
|
# https://golangci-lint.run/usage/configuration/
|
||||||
|
|
||||||
|
version: "2"
|
||||||
|
|
||||||
|
# Options for analysis running
|
||||||
|
run:
|
||||||
|
# Timeout for total work
|
||||||
|
timeout: 5m
|
||||||
|
|
||||||
|
# Skip directories (not allowed in config v2, will use issues exclude instead)
|
||||||
|
|
||||||
|
# Go version
|
||||||
|
go: "1.24"
|
||||||
|
|
||||||
|
# Include test files
|
||||||
|
tests: true
|
||||||
|
|
||||||
|
# Use Go module mode
|
||||||
|
modules-download-mode: readonly
|
||||||
|
|
||||||
|
# Output configuration
|
||||||
|
output:
|
||||||
|
# Format of output
|
||||||
|
formats:
|
||||||
|
text:
|
||||||
|
print-issued-lines: true
|
||||||
|
print-linter-name: true
|
||||||
|
|
||||||
|
# Sort results
|
||||||
|
sort-order:
|
||||||
|
- linter
|
||||||
|
- severity
|
||||||
|
- file
|
||||||
|
|
||||||
|
# Show statistics
|
||||||
|
show-stats: true
|
||||||
|
|
||||||
|
# Issues configuration
|
||||||
|
issues:
|
||||||
|
# Maximum issues count per one linter
|
||||||
|
max-issues-per-linter: 0
|
||||||
|
|
||||||
|
# Maximum count of issues with the same text
|
||||||
|
max-same-issues: 3
|
||||||
|
|
||||||
|
# Show only new issues
|
||||||
|
new: false
|
||||||
|
|
||||||
|
# Fix found issues (if linter supports)
|
||||||
|
fix: false
|
||||||
|
|
||||||
|
# Formatters configuration
|
||||||
|
formatters:
|
||||||
|
enable:
|
||||||
|
- gofmt
|
||||||
|
- goimports
|
||||||
|
- gofumpt
|
||||||
|
settings:
|
||||||
|
# gofmt settings
|
||||||
|
gofmt:
|
||||||
|
simplify: true
|
||||||
|
|
||||||
|
# goimports settings
|
||||||
|
goimports:
|
||||||
|
local-prefixes:
|
||||||
|
- github.com/kjanat/articulate-parser
|
||||||
|
|
||||||
|
# gofumpt settings
|
||||||
|
gofumpt:
|
||||||
|
module-path: github.com/kjanat/articulate-parser
|
||||||
|
extra-rules: true
|
||||||
|
|
||||||
|
# Linters configuration
|
||||||
|
linters:
|
||||||
|
# Default set of linters
|
||||||
|
default: none
|
||||||
|
|
||||||
|
# Enable specific linters
|
||||||
|
enable:
|
||||||
|
# Default/standard linters
|
||||||
|
- errcheck # Check for unchecked errors
|
||||||
|
- govet # Go vet
|
||||||
|
- ineffassign # Detect ineffectual assignments
|
||||||
|
- staticcheck # Staticcheck
|
||||||
|
- unused # Find unused code
|
||||||
|
|
||||||
|
# Code quality
|
||||||
|
- revive # Fast, configurable linter
|
||||||
|
- gocritic # Opinionated Go source code linter
|
||||||
|
- godot # Check comment periods
|
||||||
|
- godox # Detect TODO/FIXME comments
|
||||||
|
- gocognit # Cognitive complexity
|
||||||
|
- gocyclo # Cyclomatic complexity
|
||||||
|
- funlen # Function length
|
||||||
|
- maintidx # Maintainability index
|
||||||
|
|
||||||
|
# Security
|
||||||
|
- gosec # Security problems
|
||||||
|
|
||||||
|
# Performance
|
||||||
|
- prealloc # Find slice preallocation opportunities
|
||||||
|
- bodyclose # Check HTTP response body closed
|
||||||
|
|
||||||
|
# Style and formatting
|
||||||
|
- goconst # Find repeated strings
|
||||||
|
- misspell # Find misspellings
|
||||||
|
- whitespace # Find unnecessary blank lines
|
||||||
|
- unconvert # Remove unnecessary type conversions
|
||||||
|
- dupword # Check for duplicate words
|
||||||
|
|
||||||
|
# Error handling
|
||||||
|
- errorlint # Error handling improvements
|
||||||
|
- wrapcheck # Check error wrapping
|
||||||
|
|
||||||
|
# Testing
|
||||||
|
- testifylint # Testify usage
|
||||||
|
- tparallel # Detect improper t.Parallel() usage
|
||||||
|
- thelper # Detect test helpers without t.Helper()
|
||||||
|
|
||||||
|
# Best practices
|
||||||
|
- exhaustive # Check exhaustiveness of enum switches
|
||||||
|
- nolintlint # Check nolint directives
|
||||||
|
- nakedret # Find naked returns
|
||||||
|
- nilnil # Check for redundant nil checks
|
||||||
|
- noctx # Check sending HTTP requests without context
|
||||||
|
- contextcheck # Check context propagation
|
||||||
|
- asciicheck # Check for non-ASCII identifiers
|
||||||
|
- bidichk # Check for dangerous unicode sequences
|
||||||
|
- durationcheck # Check for multiplied durations
|
||||||
|
- makezero # Find slice declarations with non-zero length
|
||||||
|
- nilerr # Find code returning nil with non-nil error
|
||||||
|
- predeclared # Find code shadowing predeclared identifiers
|
||||||
|
- promlinter # Check Prometheus metrics naming
|
||||||
|
- reassign # Check reassignment of package variables
|
||||||
|
- usestdlibvars # Use variables from stdlib
|
||||||
|
- wastedassign # Find wasted assignments
|
||||||
|
|
||||||
|
# Documentation
|
||||||
|
- godoclint # Check godoc comments
|
||||||
|
|
||||||
|
# New
|
||||||
|
- modernize # Suggest simplifications using new Go features
|
||||||
|
|
||||||
|
# Exclusion rules for linters
|
||||||
|
exclusions:
|
||||||
|
rules:
|
||||||
|
# Exclude some linters from test files
|
||||||
|
- path: _test\.go
|
||||||
|
linters:
|
||||||
|
- gosec
|
||||||
|
- dupl
|
||||||
|
- errcheck
|
||||||
|
- goconst
|
||||||
|
- funlen
|
||||||
|
- goerr113
|
||||||
|
- gocognit
|
||||||
|
|
||||||
|
# Exclude benchmarks from some linters
|
||||||
|
- path: _bench_test\.go
|
||||||
|
linters:
|
||||||
|
- gosec
|
||||||
|
- dupl
|
||||||
|
- errcheck
|
||||||
|
- goconst
|
||||||
|
- funlen
|
||||||
|
- goerr113
|
||||||
|
- wrapcheck
|
||||||
|
|
||||||
|
# Exclude example tests
|
||||||
|
- path: _example_test\.go
|
||||||
|
linters:
|
||||||
|
- gosec
|
||||||
|
- errcheck
|
||||||
|
- funlen
|
||||||
|
- goerr113
|
||||||
|
- wrapcheck
|
||||||
|
- revive
|
||||||
|
|
||||||
|
# Exclude linters for main.go
|
||||||
|
- path: ^main\.go$
|
||||||
|
linters:
|
||||||
|
- forbidigo
|
||||||
|
|
||||||
|
# Exclude certain linters for generated files
|
||||||
|
- path: internal/version/version.go
|
||||||
|
linters:
|
||||||
|
- gochecknoglobals
|
||||||
|
- gochecknoinits
|
||||||
|
|
||||||
|
# Exclude var-naming for interfaces package (standard Go pattern for interface definitions)
|
||||||
|
- path: internal/interfaces/
|
||||||
|
text: "var-naming: avoid meaningless package names"
|
||||||
|
linters:
|
||||||
|
- revive
|
||||||
|
|
||||||
|
# Allow fmt.Print* in main package
|
||||||
|
- path: ^main\.go$
|
||||||
|
text: "use of fmt.Print"
|
||||||
|
linters:
|
||||||
|
- forbidigo
|
||||||
|
|
||||||
|
# Exclude common false positives
|
||||||
|
- text: "Error return value of .((os\\.)?std(out|err)\\..*|.*Close|.*Flush|os\\.Remove(All)?|.*print(f|ln)?|os\\.(Un)?Setenv). is not checked"
|
||||||
|
linters:
|
||||||
|
- errcheck
|
||||||
|
|
||||||
|
# Exclude error wrapping suggestions for well-known errors
|
||||||
|
- text: "non-wrapping format verb for fmt.Errorf"
|
||||||
|
linters:
|
||||||
|
- errorlint
|
||||||
|
|
||||||
|
# Linters settings
|
||||||
|
settings:
|
||||||
|
# errcheck settings
|
||||||
|
errcheck:
|
||||||
|
check-type-assertions: true
|
||||||
|
check-blank: false
|
||||||
|
|
||||||
|
# govet settings
|
||||||
|
govet:
|
||||||
|
enable-all: true
|
||||||
|
disable:
|
||||||
|
- fieldalignment # Too many false positives
|
||||||
|
- shadow # Can be noisy
|
||||||
|
|
||||||
|
# goconst settings
|
||||||
|
goconst:
|
||||||
|
min-len: 3
|
||||||
|
min-occurrences: 3
|
||||||
|
|
||||||
|
# godot settings
|
||||||
|
godot:
|
||||||
|
scope: toplevel
|
||||||
|
exclude:
|
||||||
|
- "^fixme:"
|
||||||
|
- "^todo:"
|
||||||
|
capital: true
|
||||||
|
period: true
|
||||||
|
|
||||||
|
# godox settings
|
||||||
|
godox:
|
||||||
|
keywords:
|
||||||
|
- TODO
|
||||||
|
- FIXME
|
||||||
|
- HACK
|
||||||
|
- BUG
|
||||||
|
- XXX
|
||||||
|
|
||||||
|
# misspell settings
|
||||||
|
misspell:
|
||||||
|
locale: US
|
||||||
|
|
||||||
|
# funlen settings
|
||||||
|
funlen:
|
||||||
|
lines: 100
|
||||||
|
statements: 50
|
||||||
|
|
||||||
|
# gocognit settings
|
||||||
|
gocognit:
|
||||||
|
min-complexity: 20
|
||||||
|
|
||||||
|
# gocyclo settings
|
||||||
|
gocyclo:
|
||||||
|
min-complexity: 15
|
||||||
|
|
||||||
|
# gocritic settings
|
||||||
|
gocritic:
|
||||||
|
enabled-tags:
|
||||||
|
- diagnostic
|
||||||
|
- style
|
||||||
|
- performance
|
||||||
|
- experimental
|
||||||
|
disabled-checks:
|
||||||
|
- ifElseChain
|
||||||
|
- singleCaseSwitch
|
||||||
|
- commentedOutCode
|
||||||
|
settings:
|
||||||
|
hugeParam:
|
||||||
|
sizeThreshold: 512
|
||||||
|
rangeValCopy:
|
||||||
|
sizeThreshold: 512
|
||||||
|
|
||||||
|
# gosec settings
|
||||||
|
gosec:
|
||||||
|
severity: medium
|
||||||
|
confidence: medium
|
||||||
|
excludes:
|
||||||
|
- G104 # Handled by errcheck
|
||||||
|
- G304 # File path provided as taint input
|
||||||
|
|
||||||
|
# revive settings
|
||||||
|
revive:
|
||||||
|
severity: warning
|
||||||
|
rules:
|
||||||
|
- name: blank-imports
|
||||||
|
- name: context-as-argument
|
||||||
|
- name: context-keys-type
|
||||||
|
- name: dot-imports
|
||||||
|
- name: empty-block
|
||||||
|
- name: error-naming
|
||||||
|
- name: error-return
|
||||||
|
- name: error-strings
|
||||||
|
- name: errorf
|
||||||
|
- name: exported
|
||||||
|
- name: if-return
|
||||||
|
- name: increment-decrement
|
||||||
|
- name: indent-error-flow
|
||||||
|
- name: package-comments
|
||||||
|
- name: range
|
||||||
|
- name: receiver-naming
|
||||||
|
- name: time-naming
|
||||||
|
- name: unexported-return
|
||||||
|
- name: var-declaration
|
||||||
|
- name: var-naming
|
||||||
|
|
||||||
|
# errorlint settings
|
||||||
|
errorlint:
|
||||||
|
errorf: true
|
||||||
|
errorf-multi: true
|
||||||
|
asserts: true
|
||||||
|
comparison: true
|
||||||
|
|
||||||
|
# wrapcheck settings
|
||||||
|
wrapcheck:
|
||||||
|
ignore-sigs:
|
||||||
|
- .Errorf(
|
||||||
|
- errors.New(
|
||||||
|
- errors.Unwrap(
|
||||||
|
- errors.Join(
|
||||||
|
- .WithMessage(
|
||||||
|
- .WithMessagef(
|
||||||
|
- .WithStack(
|
||||||
|
ignore-package-globs:
|
||||||
|
- github.com/kjanat/articulate-parser/*
|
||||||
|
|
||||||
|
# exhaustive settings
|
||||||
|
exhaustive:
|
||||||
|
check:
|
||||||
|
- switch
|
||||||
|
- map
|
||||||
|
default-signifies-exhaustive: true
|
||||||
|
|
||||||
|
# nolintlint settings
|
||||||
|
nolintlint:
|
||||||
|
allow-unused: false
|
||||||
|
require-explanation: true
|
||||||
|
require-specific: true
|
||||||
|
|
||||||
|
# stylecheck settings
|
||||||
|
staticcheck:
|
||||||
|
checks: [
|
||||||
|
"all",
|
||||||
|
"-ST1000",
|
||||||
|
"-ST1003",
|
||||||
|
"-ST1016",
|
||||||
|
"-ST1020",
|
||||||
|
"-ST1021",
|
||||||
|
"-ST1022",
|
||||||
|
]
|
||||||
|
|
||||||
|
# maintidx settings
|
||||||
|
maintidx:
|
||||||
|
under: 20
|
||||||
|
|
||||||
|
# testifylint settings
|
||||||
|
testifylint:
|
||||||
|
enable-all: true
|
||||||
|
disable:
|
||||||
|
- float-compare
|
||||||
|
|
||||||
|
# thelper settings
|
||||||
|
thelper:
|
||||||
|
test:
|
||||||
|
first: true
|
||||||
|
name: true
|
||||||
|
begin: true
|
||||||
|
benchmark:
|
||||||
|
first: true
|
||||||
|
name: true
|
||||||
|
begin: true
|
||||||
|
|
||||||
|
# Severity rules
|
||||||
|
severity:
|
||||||
|
default: warning
|
||||||
|
rules:
|
||||||
|
- linters:
|
||||||
|
- gosec
|
||||||
|
severity: error
|
||||||
|
- linters:
|
||||||
|
- errcheck
|
||||||
|
- staticcheck
|
||||||
|
severity: error
|
||||||
|
- linters:
|
||||||
|
- godox
|
||||||
|
severity: info
|
||||||
75
.pre-commit-config.yaml
Normal file
75
.pre-commit-config.yaml
Normal file
@ -0,0 +1,75 @@
|
|||||||
|
# See https://pre-commit.com for more information
|
||||||
|
# See https://pre-commit.com/hooks.html for more hooks
|
||||||
|
repos:
|
||||||
|
- repo: https://github.com/pre-commit/pre-commit-hooks
|
||||||
|
rev: v6.0.0
|
||||||
|
hooks:
|
||||||
|
# File quality
|
||||||
|
- id: trailing-whitespace
|
||||||
|
exclude: '^\.github/ISSUE_TEMPLATE/.*\.yml$'
|
||||||
|
- id: end-of-file-fixer
|
||||||
|
- id: mixed-line-ending
|
||||||
|
args: ["--fix=lf"]
|
||||||
|
|
||||||
|
# File validation
|
||||||
|
- id: check-yaml
|
||||||
|
- id: check-json
|
||||||
|
- id: check-toml
|
||||||
|
|
||||||
|
# Security
|
||||||
|
- id: detect-private-key
|
||||||
|
|
||||||
|
# Git safety
|
||||||
|
- id: check-merge-conflict
|
||||||
|
- id: check-case-conflict
|
||||||
|
- id: no-commit-to-branch
|
||||||
|
args: ["--branch=master", "--branch=main"]
|
||||||
|
|
||||||
|
# File structure
|
||||||
|
- id: check-added-large-files
|
||||||
|
- id: check-symlinks
|
||||||
|
- id: check-executables-have-shebangs
|
||||||
|
- repo: local
|
||||||
|
hooks:
|
||||||
|
- id: actionlint
|
||||||
|
name: Lint GitHub Actions workflow files
|
||||||
|
description: Runs actionlint to lint GitHub Actions workflow files
|
||||||
|
language: golang
|
||||||
|
types: ["yaml"]
|
||||||
|
files: ^\.github/workflows/
|
||||||
|
entry: actionlint
|
||||||
|
minimum_pre_commit_version: 3.0.0
|
||||||
|
- repo: https://github.com/golangci/golangci-lint
|
||||||
|
rev: v2.7.2
|
||||||
|
hooks:
|
||||||
|
- id: golangci-lint
|
||||||
|
name: golangci-lint
|
||||||
|
description: Fast linters runner for Go. Note that only modified files are linted, so linters like 'unused' that need to scan all files won't work as expected.
|
||||||
|
entry: golangci-lint run --new-from-rev HEAD --fix
|
||||||
|
types: [go]
|
||||||
|
language: golang
|
||||||
|
require_serial: true
|
||||||
|
pass_filenames: false
|
||||||
|
# - id: golangci-lint-full
|
||||||
|
# name: golangci-lint-full
|
||||||
|
# description: Fast linters runner for Go. Runs on all files in the module. Use this hook if you use pre-commit in CI.
|
||||||
|
# entry: golangci-lint run --fix
|
||||||
|
# types: [go]
|
||||||
|
# language: golang
|
||||||
|
# require_serial: true
|
||||||
|
# pass_filenames: false
|
||||||
|
- id: golangci-lint-fmt
|
||||||
|
name: golangci-lint-fmt
|
||||||
|
description: Fast linters runner for Go. Formats all files in the repo.
|
||||||
|
entry: golangci-lint fmt
|
||||||
|
types: [go]
|
||||||
|
language: golang
|
||||||
|
require_serial: true
|
||||||
|
pass_filenames: false
|
||||||
|
- id: golangci-lint-config-verify
|
||||||
|
name: golangci-lint-config-verify
|
||||||
|
description: Verifies the configuration file
|
||||||
|
entry: golangci-lint config verify
|
||||||
|
files: '\.golangci\.(?:yml|yaml|toml|json)'
|
||||||
|
language: golang
|
||||||
|
pass_filenames: false
|
||||||
183
AGENTS.md
Normal file
183
AGENTS.md
Normal file
@ -0,0 +1,183 @@
|
|||||||
|
# Agent Guidelines for articulate-parser
|
||||||
|
|
||||||
|
A Go CLI tool that parses Articulate Rise courses from URLs or local JSON files and exports them to Markdown, HTML, or DOCX formats.
|
||||||
|
|
||||||
|
## Repository Info
|
||||||
|
|
||||||
|
- **GitHub**: https://github.com/kjanat/articulate-parser
|
||||||
|
- **Default branch**: `master` (not `main`)
|
||||||
|
|
||||||
|
## Build/Test Commands
|
||||||
|
|
||||||
|
### Primary Commands (using Taskfile)
|
||||||
|
|
||||||
|
```bash
|
||||||
|
task build # Build binary to bin/articulate-parser
|
||||||
|
task test # Run all tests with race detection
|
||||||
|
task lint # Run all linters (vet, fmt, staticcheck, golangci-lint)
|
||||||
|
task fmt # Format all Go files
|
||||||
|
task ci # Full CI pipeline: deps, lint, test with coverage, build
|
||||||
|
task qa # Quick QA: fmt + lint + test
|
||||||
|
```
|
||||||
|
|
||||||
|
### Direct Go Commands
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Build
|
||||||
|
go build -o bin/articulate-parser main.go
|
||||||
|
|
||||||
|
# Run all tests
|
||||||
|
go test -race -timeout 5m ./...
|
||||||
|
|
||||||
|
# Run single test by name
|
||||||
|
go test -v -race -run ^TestMarkdownExporter_Export$ ./internal/exporters
|
||||||
|
|
||||||
|
# Run tests in specific package
|
||||||
|
go test -v -race ./internal/services
|
||||||
|
|
||||||
|
# Run tests matching pattern
|
||||||
|
go test -v -race -run "TestParser" ./...
|
||||||
|
|
||||||
|
# Test with coverage
|
||||||
|
go test -race -coverprofile=coverage/coverage.out -covermode=atomic ./...
|
||||||
|
go tool cover -html=coverage/coverage.out -o coverage/coverage.html
|
||||||
|
|
||||||
|
# Benchmarks
|
||||||
|
go test -bench=. -benchmem ./...
|
||||||
|
go test -bench=BenchmarkMarkdownExporter ./internal/exporters
|
||||||
|
```
|
||||||
|
|
||||||
|
### Security & Auditing
|
||||||
|
|
||||||
|
```bash
|
||||||
|
task security:check # Run gosec security scanner
|
||||||
|
task security:audit # Run govulncheck for vulnerabilities
|
||||||
|
```
|
||||||
|
|
||||||
|
## Code Style Guidelines
|
||||||
|
|
||||||
|
### Imports
|
||||||
|
|
||||||
|
- Use `goimports` with local prefix: `github.com/kjanat/articulate-parser`
|
||||||
|
- Order: stdlib, blank line, external packages, blank line, internal packages
|
||||||
|
|
||||||
|
```go
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/fumiama/go-docx"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/interfaces"
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
### Formatting
|
||||||
|
|
||||||
|
- Use `gofmt -s` (simplify) and `gofumpt` with extra rules
|
||||||
|
- Function length: max 100 lines, 50 statements
|
||||||
|
- Cyclomatic complexity: max 15; Cognitive complexity: max 20
|
||||||
|
|
||||||
|
### Types & Naming
|
||||||
|
|
||||||
|
- Use interface-based design (see `internal/interfaces/`)
|
||||||
|
- Exported types/functions require godoc comments ending with period
|
||||||
|
- Use descriptive names: `ArticulateParser`, `MarkdownExporter`
|
||||||
|
- Receiver names: short (1-2 chars), consistent per type
|
||||||
|
|
||||||
|
### Error Handling
|
||||||
|
|
||||||
|
- Always wrap errors with context: `fmt.Errorf("operation failed: %w", err)`
|
||||||
|
- Use `%w` verb for error wrapping to preserve error chain
|
||||||
|
- Check all error returns (enforced by `errcheck`)
|
||||||
|
- Document error handling rationale in defer blocks when ignoring close errors
|
||||||
|
|
||||||
|
```go
|
||||||
|
// Good: Error wrapping with context
|
||||||
|
if err := json.Unmarshal(body, &course); err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to unmarshal JSON: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Good: Documented defer with error handling
|
||||||
|
defer func() {
|
||||||
|
if err := resp.Body.Close(); err != nil {
|
||||||
|
p.Logger.Warn("failed to close response body", "error", err)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
```
|
||||||
|
|
||||||
|
### Comments
|
||||||
|
|
||||||
|
- All exported types/functions require godoc comments
|
||||||
|
- End sentences with periods (`godot` linter enforced)
|
||||||
|
- Mark known issues with TODO/FIXME/HACK/BUG/XXX
|
||||||
|
|
||||||
|
### Security
|
||||||
|
|
||||||
|
- Use `#nosec` with justification for deliberate security exceptions
|
||||||
|
- G304: File paths from CLI args; G306: Export file permissions
|
||||||
|
|
||||||
|
```go
|
||||||
|
// #nosec G304 - File path provided by user via CLI argument
|
||||||
|
data, err := os.ReadFile(filePath)
|
||||||
|
```
|
||||||
|
|
||||||
|
### Testing
|
||||||
|
|
||||||
|
- Enable race detection: `-race` flag always
|
||||||
|
- Use table-driven tests where applicable
|
||||||
|
- Mark test helpers with `t.Helper()`
|
||||||
|
- Use `t.TempDir()` for temporary files
|
||||||
|
- Benchmarks in `*_bench_test.go`, examples in `*_example_test.go`
|
||||||
|
- Test naming: `Test<Type>_<Method>` or `Test<Function>`
|
||||||
|
|
||||||
|
```go
|
||||||
|
func TestMarkdownExporter_ProcessItemToMarkdown_AllTypes(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
name, itemType, expectedText string
|
||||||
|
}{
|
||||||
|
{"text item", "text", ""},
|
||||||
|
{"divider item", "divider", "---"},
|
||||||
|
}
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
// test implementation
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### Dependencies
|
||||||
|
|
||||||
|
- Minimal external dependencies (go-docx, golang.org/x/net, golang.org/x/text)
|
||||||
|
- Run `task deps:tidy` after adding/removing dependencies
|
||||||
|
- CGO disabled by default (`CGO_ENABLED=0`)
|
||||||
|
|
||||||
|
## Project Structure
|
||||||
|
|
||||||
|
```
|
||||||
|
articulate-parser/
|
||||||
|
internal/
|
||||||
|
config/ # Configuration loading
|
||||||
|
exporters/ # Export implementations (markdown, html, docx)
|
||||||
|
interfaces/ # Core interfaces (Exporter, CourseParser, Logger)
|
||||||
|
models/ # Data models (Course, Lesson, Item, Media)
|
||||||
|
services/ # Core services (parser, html cleaner, app, logger)
|
||||||
|
version/ # Version information
|
||||||
|
main.go # Application entry point
|
||||||
|
```
|
||||||
|
|
||||||
|
## Common Patterns
|
||||||
|
|
||||||
|
### Creating a new exporter
|
||||||
|
|
||||||
|
1. Implement `interfaces.Exporter` interface
|
||||||
|
2. Add factory method to `internal/exporters/factory.go`
|
||||||
|
3. Register format in `NewFactory()`
|
||||||
|
4. Add tests following existing patterns
|
||||||
|
|
||||||
|
### Adding configuration options
|
||||||
|
|
||||||
|
1. Add field to `Config` struct in `internal/config/config.go`
|
||||||
|
2. Load from environment variable with sensible default
|
||||||
|
3. Document in config struct comments
|
||||||
22
DOCKER.md
22
DOCKER.md
@ -49,17 +49,17 @@ docker run --rm ghcr.io/kjanat/articulate-parser:latest --version
|
|||||||
|
|
||||||
## Available Tags
|
## Available Tags
|
||||||
|
|
||||||
- `latest` - Latest stable release
|
- `latest` - Latest stable release
|
||||||
- `v1.x.x` - Specific version tags
|
- `v1.x.x` - Specific version tags
|
||||||
- `main` - Latest development build
|
- `main` - Latest development build
|
||||||
|
|
||||||
## Image Details
|
## Image Details
|
||||||
|
|
||||||
- **Base Image**: `scratch` (minimal attack surface)
|
- **Base Image**: `scratch` (minimal attack surface)
|
||||||
- **Architecture**: Multi-arch support (amd64, arm64)
|
- **Architecture**: Multi-arch support (amd64, arm64)
|
||||||
- **Size**: < 10MB (optimized binary)
|
- **Size**: < 10MB (optimized binary)
|
||||||
- **Security**: Runs as non-root user
|
- **Security**: Runs as non-root user
|
||||||
- **Features**: SBOM and provenance attestation included
|
- **Features**: SBOM and provenance attestation included
|
||||||
|
|
||||||
## Development
|
## Development
|
||||||
|
|
||||||
@ -77,6 +77,6 @@ docker-compose up --build
|
|||||||
|
|
||||||
## Repository
|
## Repository
|
||||||
|
|
||||||
- **Source**: [github.com/kjanat/articulate-parser](https://github.com/kjanat/articulate-parser)
|
- **Source**: [github.com/kjanat/articulate-parser](https://github.com/kjanat/articulate-parser)
|
||||||
- **Issues**: [Report bugs or request features](https://github.com/kjanat/articulate-parser/issues)
|
- **Issues**: [Report bugs or request features](https://github.com/kjanat/articulate-parser/issues)
|
||||||
- **License**: See repository for license details
|
- **License**: See repository for license details
|
||||||
|
|||||||
@ -1,5 +1,5 @@
|
|||||||
# Build stage
|
# Build stage
|
||||||
FROM golang:1.24-alpine AS builder
|
FROM golang:1.25-alpine AS builder
|
||||||
|
|
||||||
# Install git and ca-certificates (needed for fetching dependencies and HTTPS)
|
# Install git and ca-certificates (needed for fetching dependencies and HTTPS)
|
||||||
RUN apk add --no-cache git ca-certificates tzdata file
|
RUN apk add --no-cache git ca-certificates tzdata file
|
||||||
|
|||||||
@ -2,7 +2,7 @@
|
|||||||
# Uses Alpine instead of scratch for debugging
|
# Uses Alpine instead of scratch for debugging
|
||||||
|
|
||||||
# Build stage - same as production
|
# Build stage - same as production
|
||||||
FROM golang:1.24-alpine AS builder
|
FROM golang:1.25-alpine AS builder
|
||||||
|
|
||||||
# Install git and ca-certificates (needed for fetching dependencies and HTTPS)
|
# Install git and ca-certificates (needed for fetching dependencies and HTTPS)
|
||||||
RUN apk add --no-cache git ca-certificates tzdata file
|
RUN apk add --no-cache git ca-certificates tzdata file
|
||||||
@ -49,7 +49,7 @@ RUN CGO_ENABLED=0 GOOS=${TARGETOS} GOARCH=${TARGETARCH} go build \
|
|||||||
RUN file /app/articulate-parser || echo "file command not available"
|
RUN file /app/articulate-parser || echo "file command not available"
|
||||||
|
|
||||||
# Development stage - uses Alpine for shell access
|
# Development stage - uses Alpine for shell access
|
||||||
FROM alpine:3.21.3
|
FROM alpine:3
|
||||||
|
|
||||||
# Install minimal dependencies
|
# Install minimal dependencies
|
||||||
RUN apk add --no-cache ca-certificates tzdata
|
RUN apk add --no-cache ca-certificates tzdata
|
||||||
|
|||||||
198
README.md
198
README.md
@ -78,32 +78,32 @@ flowchart TD
|
|||||||
|
|
||||||
The system follows **Clean Architecture** principles with clear separation of concerns:
|
The system follows **Clean Architecture** principles with clear separation of concerns:
|
||||||
|
|
||||||
- **🎯 Entry Point**: Command-line interface handles user input and coordinates operations
|
- **🎯 Entry Point**: Command-line interface handles user input and coordinates operations
|
||||||
- **🏗️ Application Layer**: Core business logic with dependency injection
|
- **🏗️ Application Layer**: Core business logic with dependency injection
|
||||||
- **📋 Interface Layer**: Contracts defining behavior without implementation details
|
- **📋 Interface Layer**: Contracts defining behavior without implementation details
|
||||||
- **🔧 Service Layer**: Concrete implementations of parsing and utility services
|
- **🔧 Service Layer**: Concrete implementations of parsing and utility services
|
||||||
- **📤 Export Layer**: Factory pattern for format-specific exporters
|
- **📤 Export Layer**: Factory pattern for format-specific exporters
|
||||||
- **📊 Data Layer**: Domain models representing course structure
|
- **📊 Data Layer**: Domain models representing course structure
|
||||||
|
|
||||||
## Features
|
## Features
|
||||||
|
|
||||||
- Parse Articulate Rise JSON data from URLs or local files
|
- Parse Articulate Rise JSON data from URLs or local files
|
||||||
- Export to Markdown (.md) format
|
- Export to Markdown (.md) format
|
||||||
- Export to HTML (.html) format with professional styling
|
- Export to HTML (.html) format with professional styling
|
||||||
- Export to Word Document (.docx) format
|
- Export to Word Document (.docx) format
|
||||||
- Support for various content types:
|
- Support for various content types:
|
||||||
- Text content with headings and paragraphs
|
- Text content with headings and paragraphs
|
||||||
- Lists and bullet points
|
- Lists and bullet points
|
||||||
- Multimedia content (videos and images)
|
- Multimedia content (videos and images)
|
||||||
- Knowledge checks and quizzes
|
- Knowledge checks and quizzes
|
||||||
- Interactive content (flashcards)
|
- Interactive content (flashcards)
|
||||||
- Course structure and metadata
|
- Course structure and metadata
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
### Prerequisites
|
### Prerequisites
|
||||||
|
|
||||||
- Go, I don't know the version, but I have [][gomod] configured right now, and it works, see the [CI][Build] workflow where it is tested.
|
- Go, I don't know the version, but I have [][gomod] configured right now, and it works, see the [CI][Build] workflow where it is tested.
|
||||||
|
|
||||||
### Install from source
|
### Install from source
|
||||||
|
|
||||||
@ -124,7 +124,7 @@ go install github.com/kjanat/articulate-parser@latest
|
|||||||
|
|
||||||
The parser uses the following external library:
|
The parser uses the following external library:
|
||||||
|
|
||||||
- `github.com/fumiama/go-docx` - For creating Word documents (MIT license)
|
- `github.com/fumiama/go-docx` - For creating Word documents (MIT license)
|
||||||
|
|
||||||
## Testing
|
## Testing
|
||||||
|
|
||||||
@ -164,25 +164,25 @@ go run main.go <input_uri_or_file> <output_format> [output_path]
|
|||||||
|
|
||||||
#### Examples
|
#### Examples
|
||||||
|
|
||||||
1. **Parse from URL and export to Markdown:**
|
1. **Parse from URL and export to Markdown:**
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
go run main.go "https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/" md
|
go run main.go "https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/" md
|
||||||
```
|
```
|
||||||
|
|
||||||
2. **Parse from local file and export to Word:**
|
2. **Parse from local file and export to Word:**
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
go run main.go "articulate-sample.json" docx "my-course.docx"
|
go run main.go "articulate-sample.json" docx "my-course.docx"
|
||||||
```
|
```
|
||||||
|
|
||||||
3. **Parse from local file and export to HTML:**
|
3. **Parse from local file and export to HTML:**
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
go run main.go "articulate-sample.json" html "output.html"
|
go run main.go "articulate-sample.json" html "output.html"
|
||||||
```
|
```
|
||||||
|
|
||||||
4. **Parse from local file and export to Markdown:**
|
4. **Parse from local file and export to Markdown:**
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
go run main.go "articulate-sample.json" md "output.md"
|
go run main.go "articulate-sample.json" md "output.md"
|
||||||
@ -225,14 +225,14 @@ docker run --rm ghcr.io/kjanat/articulate-parser:latest --help
|
|||||||
|
|
||||||
### Available Tags
|
### Available Tags
|
||||||
|
|
||||||
| Tag | Description | Use Case |
|
| Tag | Description | Use Case |
|
||||||
|-----|-------------|----------|
|
| --------------------- | ------------------------------------------- | ---------------------- |
|
||||||
| `latest` | Latest stable release from master branch | Production use |
|
| `latest` | Latest stable release from master branch | Production use |
|
||||||
| `edge` | Latest development build from master branch | Testing new features |
|
| `edge` | Latest development build from master branch | Testing new features |
|
||||||
| `v1.x.x` | Specific version releases | Production pinning |
|
| `v1.x.x` | Specific version releases | Production pinning |
|
||||||
| `develop` | Development branch builds | Development/testing |
|
| `develop` | Development branch builds | Development/testing |
|
||||||
| `feature/docker-ghcr` | Feature branch builds | Feature testing |
|
| `feature/docker-ghcr` | Feature branch builds | Feature testing |
|
||||||
| `master` | Latest master branch build | Continuous integration |
|
| `master` | Latest master branch build | Continuous integration |
|
||||||
|
|
||||||
### Usage Examples
|
### Usage Examples
|
||||||
|
|
||||||
@ -313,11 +313,11 @@ docker build --build-arg VERSION=local --build-arg BUILD_TIME=$(date -u +%Y-%m-%
|
|||||||
|
|
||||||
The Docker image supports the following build-time arguments:
|
The Docker image supports the following build-time arguments:
|
||||||
|
|
||||||
| Argument | Description | Default |
|
| Argument | Description | Default |
|
||||||
|----------|-------------|---------|
|
| ------------ | ------------------------------------- | -------------- |
|
||||||
| `VERSION` | Version string embedded in the binary | `dev` |
|
| `VERSION` | Version string embedded in the binary | `dev` |
|
||||||
| `BUILD_TIME` | Build timestamp | Current time |
|
| `BUILD_TIME` | Build timestamp | Current time |
|
||||||
| `GIT_COMMIT` | Git commit hash | Current commit |
|
| `GIT_COMMIT` | Git commit hash | Current commit |
|
||||||
|
|
||||||
### Docker Security
|
### Docker Security
|
||||||
|
|
||||||
@ -332,88 +332,88 @@ The Docker image supports the following build-time arguments:
|
|||||||
|
|
||||||
The project maintains high code quality standards:
|
The project maintains high code quality standards:
|
||||||
|
|
||||||
- Cyclomatic complexity ≤ 15 (checked with [gocyclo](https://github.com/fzipp/gocyclo))
|
- Cyclomatic complexity ≤ 15 (checked with [gocyclo](https://github.com/fzipp/gocyclo))
|
||||||
- Race condition detection enabled
|
- Race condition detection enabled
|
||||||
- Comprehensive test coverage
|
- Comprehensive test coverage
|
||||||
- Code formatting with `gofmt`
|
- Code formatting with `gofmt`
|
||||||
- Static analysis with `go vet`
|
- Static analysis with `go vet`
|
||||||
|
|
||||||
### Contributing
|
### Contributing
|
||||||
|
|
||||||
1. Fork the repository
|
1. Fork the repository
|
||||||
2. Create a feature branch
|
2. Create a feature branch
|
||||||
3. Make your changes
|
3. Make your changes
|
||||||
4. Run tests: `go test ./...`
|
4. Run tests: `go test ./...`
|
||||||
5. Submit a pull request
|
5. Submit a pull request
|
||||||
|
|
||||||
## Output Formats
|
## Output Formats
|
||||||
|
|
||||||
### Markdown (`.md`)
|
### Markdown (`.md`)
|
||||||
|
|
||||||
- Hierarchical structure with proper heading levels
|
- Hierarchical structure with proper heading levels
|
||||||
- Clean text content with HTML tags removed
|
- Clean text content with HTML tags removed
|
||||||
- Lists and bullet points preserved
|
- Lists and bullet points preserved
|
||||||
- Quiz questions with correct answers marked
|
- Quiz questions with correct answers marked
|
||||||
- Media references included
|
- Media references included
|
||||||
- Course metadata at the top
|
- Course metadata at the top
|
||||||
|
|
||||||
### HTML (`.html`)
|
### HTML (`.html`)
|
||||||
|
|
||||||
- Professional styling with embedded CSS
|
- Professional styling with embedded CSS
|
||||||
- Interactive and visually appealing layout
|
- Interactive and visually appealing layout
|
||||||
- Proper HTML structure with semantic elements
|
- Proper HTML structure with semantic elements
|
||||||
- Responsive design for different screen sizes
|
- Responsive design for different screen sizes
|
||||||
- All content types beautifully formatted
|
- All content types beautifully formatted
|
||||||
- Maintains course hierarchy and organization
|
- Maintains course hierarchy and organization
|
||||||
|
|
||||||
### Word Document (`.docx`)
|
### Word Document (`.docx`)
|
||||||
|
|
||||||
- Professional document formatting
|
- Professional document formatting
|
||||||
- Bold headings and proper typography
|
- Bold headings and proper typography
|
||||||
- Bulleted lists
|
- Bulleted lists
|
||||||
- Quiz questions with answers
|
- Quiz questions with answers
|
||||||
- Media content references
|
- Media content references
|
||||||
- Maintains course structure
|
- Maintains course structure
|
||||||
|
|
||||||
## Supported Content Types
|
## Supported Content Types
|
||||||
|
|
||||||
The parser handles the following Articulate Rise content types:
|
The parser handles the following Articulate Rise content types:
|
||||||
|
|
||||||
- **Text blocks**: Headings and paragraphs
|
- **Text blocks**: Headings and paragraphs
|
||||||
- **Lists**: Bullet points and numbered lists
|
- **Lists**: Bullet points and numbered lists
|
||||||
- **Multimedia**: Videos and images (references only)
|
- **Multimedia**: Videos and images (references only)
|
||||||
- **Knowledge Checks**: Multiple choice, multiple response, fill-in-the-blank, matching
|
- **Knowledge Checks**: Multiple choice, multiple response, fill-in-the-blank, matching
|
||||||
- **Interactive Content**: Flashcards and interactive scenarios
|
- **Interactive Content**: Flashcards and interactive scenarios
|
||||||
- **Dividers**: Section breaks
|
- **Dividers**: Section breaks
|
||||||
- **Sections**: Course organization
|
- **Sections**: Course organization
|
||||||
|
|
||||||
## Data Structure
|
## Data Structure
|
||||||
|
|
||||||
The parser works with the standard Articulate Rise JSON format which includes:
|
The parser works with the standard Articulate Rise JSON format which includes:
|
||||||
|
|
||||||
- Course metadata (title, description, settings)
|
- Course metadata (title, description, settings)
|
||||||
- Lesson structure
|
- Lesson structure
|
||||||
- Content items with various types
|
- Content items with various types
|
||||||
- Media references
|
- Media references
|
||||||
- Quiz/assessment data
|
- Quiz/assessment data
|
||||||
- Styling and layout information
|
- Styling and layout information
|
||||||
|
|
||||||
## URL Pattern Recognition
|
## URL Pattern Recognition
|
||||||
|
|
||||||
The parser automatically extracts share IDs from Articulate Rise URLs:
|
The parser automatically extracts share IDs from Articulate Rise URLs:
|
||||||
|
|
||||||
- Input: `https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/`
|
- Input: `https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/`
|
||||||
- API URL: `https://rise.articulate.com/api/rise-runtime/boot/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO`
|
- API URL: `https://rise.articulate.com/api/rise-runtime/boot/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO`
|
||||||
|
|
||||||
## Error Handling
|
## Error Handling
|
||||||
|
|
||||||
The parser includes error handling for:
|
The parser includes error handling for:
|
||||||
|
|
||||||
- Invalid URLs or share IDs
|
- Invalid URLs or share IDs
|
||||||
- Network connection issues
|
- Network connection issues
|
||||||
- Malformed JSON data
|
- Malformed JSON data
|
||||||
- File I/O errors
|
- File I/O errors
|
||||||
- Unsupported content types
|
- Unsupported content types
|
||||||
|
|
||||||
<!-- ## Code coverage
|
<!-- ## Code coverage
|
||||||
|
|
||||||
@ -425,28 +425,28 @@ The parser includes error handling for:
|
|||||||
|
|
||||||
## Limitations
|
## Limitations
|
||||||
|
|
||||||
- Media files (videos, images) are referenced but not downloaded
|
- Media files (videos, images) are referenced but not downloaded
|
||||||
- Complex interactive elements may be simplified in export
|
- Complex interactive elements may be simplified in export
|
||||||
- Styling and visual formatting is not preserved
|
- Styling and visual formatting is not preserved
|
||||||
- Assessment logic and interactivity is lost in static exports
|
- Assessment logic and interactivity is lost in static exports
|
||||||
|
|
||||||
## Performance
|
## Performance
|
||||||
|
|
||||||
- Lightweight with minimal dependencies
|
- Lightweight with minimal dependencies
|
||||||
- Fast JSON parsing and export
|
- Fast JSON parsing and export
|
||||||
- Memory efficient processing
|
- Memory efficient processing
|
||||||
- No external license requirements
|
- No external license requirements
|
||||||
|
|
||||||
## Future Enhancements
|
## Future Enhancements
|
||||||
|
|
||||||
Potential improvements could include:
|
Potential improvements could include:
|
||||||
|
|
||||||
- [ ] PDF export support
|
- [ ] PDF export support
|
||||||
- [ ] Media file downloading
|
- [ ] Media file downloading
|
||||||
- [x] ~~HTML export with preserved styling~~
|
- [x] ~~HTML export with preserved styling~~
|
||||||
- [ ] SCORM package support
|
- [ ] SCORM package support
|
||||||
- [ ] Batch processing capabilities
|
- [ ] Batch processing capabilities
|
||||||
- [ ] Custom template support
|
- [ ] Custom template support
|
||||||
|
|
||||||
## License
|
## License
|
||||||
|
|
||||||
@ -460,7 +460,9 @@ This is a utility tool for educational content conversion. Please ensure you hav
|
|||||||
[Go report]: https://goreportcard.com/report/github.com/kjanat/articulate-parser
|
[Go report]: https://goreportcard.com/report/github.com/kjanat/articulate-parser
|
||||||
[gomod]: go.mod
|
[gomod]: go.mod
|
||||||
[Issues]: https://github.com/kjanat/articulate-parser/issues
|
[Issues]: https://github.com/kjanat/articulate-parser/issues
|
||||||
|
|
||||||
<!-- [Latest release]: https://github.com/kjanat/articulate-parser/releases/latest -->
|
<!-- [Latest release]: https://github.com/kjanat/articulate-parser/releases/latest -->
|
||||||
|
|
||||||
[MIT License]: LICENSE
|
[MIT License]: LICENSE
|
||||||
[Package documentation]: https://godoc.org/github.com/kjanat/articulate-parser
|
[Package documentation]: https://godoc.org/github.com/kjanat/articulate-parser
|
||||||
[Tags]: https://github.com/kjanat/articulate-parser/tags
|
[Tags]: https://github.com/kjanat/articulate-parser/tags
|
||||||
|
|||||||
610
Taskfile.yml
Normal file
610
Taskfile.yml
Normal file
@ -0,0 +1,610 @@
|
|||||||
|
# yaml-language-server: $schema=https://taskfile.dev/schema.json
|
||||||
|
# Articulate Parser - Task Automation
|
||||||
|
# https://taskfile.dev
|
||||||
|
version: "3"
|
||||||
|
|
||||||
|
# Global output settings
|
||||||
|
output: prefixed
|
||||||
|
|
||||||
|
# Shell settings (only applied on Unix-like systems)
|
||||||
|
# Note: These are ignored on Windows where PowerShell/cmd is used
|
||||||
|
set: [errexit, pipefail]
|
||||||
|
shopt: [globstar]
|
||||||
|
|
||||||
|
# Watch mode interval
|
||||||
|
interval: 500ms
|
||||||
|
|
||||||
|
# Global variables
|
||||||
|
vars:
|
||||||
|
APP_NAME: articulate-parser
|
||||||
|
MAIN_FILE: main.go
|
||||||
|
OUTPUT_DIR: bin
|
||||||
|
COVERAGE_DIR: coverage
|
||||||
|
TEST_TIMEOUT: 5m
|
||||||
|
|
||||||
|
# Version info
|
||||||
|
VERSION:
|
||||||
|
sh: git describe --tags --always --dirty 2>/dev/null || echo "dev"
|
||||||
|
GIT_COMMIT:
|
||||||
|
sh: git rev-parse --short HEAD 2>/dev/null || echo "unknown"
|
||||||
|
BUILD_TIME: '{{now | date "2006-01-02T15:04:05Z07:00"}}'
|
||||||
|
|
||||||
|
# Go settings
|
||||||
|
CGO_ENABLED: 0
|
||||||
|
GO_FLAGS: -v
|
||||||
|
LDFLAGS: >-
|
||||||
|
-s -w
|
||||||
|
-X github.com/kjanat/articulate-parser/internal/version.Version={{.VERSION}}
|
||||||
|
-X github.com/kjanat/articulate-parser/internal/version.BuildTime={{.BUILD_TIME}}
|
||||||
|
-X github.com/kjanat/articulate-parser/internal/version.GitCommit={{.GIT_COMMIT}}
|
||||||
|
|
||||||
|
# Platform detection (using Task built-in variables)
|
||||||
|
GOOS:
|
||||||
|
sh: go env GOOS
|
||||||
|
GOARCH:
|
||||||
|
sh: go env GOARCH
|
||||||
|
EXE_EXT: '{{if eq OS "windows"}}.exe{{end}}'
|
||||||
|
|
||||||
|
# Environment variables
|
||||||
|
env:
|
||||||
|
CGO_ENABLED: "{{.CGO_ENABLED}}"
|
||||||
|
GO111MODULE: on
|
||||||
|
|
||||||
|
# Load .env files if present
|
||||||
|
dotenv: [".env", ".env.local"]
|
||||||
|
|
||||||
|
# Task definitions
|
||||||
|
tasks:
|
||||||
|
# Default task - show help
|
||||||
|
default:
|
||||||
|
desc: Show available tasks
|
||||||
|
cmds:
|
||||||
|
- task --list
|
||||||
|
silent: true
|
||||||
|
|
||||||
|
# Development tasks
|
||||||
|
dev:
|
||||||
|
desc: Run the application in development mode (with hot reload)
|
||||||
|
aliases: [run, start]
|
||||||
|
interactive: true
|
||||||
|
watch: true
|
||||||
|
sources:
|
||||||
|
- "**/*.go"
|
||||||
|
- go.mod
|
||||||
|
- go.sum
|
||||||
|
cmds:
|
||||||
|
- task: build
|
||||||
|
- "{{.OUTPUT_DIR}}/{{.APP_NAME}}{{.EXE_EXT}} --help"
|
||||||
|
|
||||||
|
# Build tasks
|
||||||
|
build:
|
||||||
|
desc: Build the application binary
|
||||||
|
aliases: [b]
|
||||||
|
deps: [clean-bin]
|
||||||
|
sources:
|
||||||
|
- "**/*.go"
|
||||||
|
- go.mod
|
||||||
|
- go.sum
|
||||||
|
generates:
|
||||||
|
- "{{.OUTPUT_DIR}}/{{.APP_NAME}}{{.EXE_EXT}}"
|
||||||
|
cmds:
|
||||||
|
- task: mkdir
|
||||||
|
vars: { DIR: "{{.OUTPUT_DIR}}" }
|
||||||
|
- go build {{.GO_FLAGS}} -ldflags="{{.LDFLAGS}}" -o {{.OUTPUT_DIR}}/{{.APP_NAME}}{{.EXE_EXT}} {{.MAIN_FILE}}
|
||||||
|
method: checksum
|
||||||
|
|
||||||
|
build:all:
|
||||||
|
desc: Build binaries for all major platforms
|
||||||
|
aliases: [build-all, cross-compile]
|
||||||
|
deps: [clean-bin]
|
||||||
|
cmds:
|
||||||
|
- task: mkdir
|
||||||
|
vars: { DIR: "{{.OUTPUT_DIR}}" }
|
||||||
|
- for:
|
||||||
|
matrix:
|
||||||
|
GOOS: [linux, darwin, windows]
|
||||||
|
GOARCH: [amd64, arm64]
|
||||||
|
task: build:platform
|
||||||
|
vars:
|
||||||
|
TARGET_GOOS: "{{.ITEM.GOOS}}"
|
||||||
|
TARGET_GOARCH: "{{.ITEM.GOARCH}}"
|
||||||
|
- echo "Built binaries for all platforms in {{.OUTPUT_DIR}}/"
|
||||||
|
|
||||||
|
build:platform:
|
||||||
|
internal: true
|
||||||
|
vars:
|
||||||
|
TARGET_EXT: '{{if eq .TARGET_GOOS "windows"}}.exe{{end}}'
|
||||||
|
OUTPUT_FILE: "{{.OUTPUT_DIR}}/{{.APP_NAME}}-{{.TARGET_GOOS}}-{{.TARGET_GOARCH}}{{.TARGET_EXT}}"
|
||||||
|
env:
|
||||||
|
GOOS: "{{.TARGET_GOOS}}"
|
||||||
|
GOARCH: "{{.TARGET_GOARCH}}"
|
||||||
|
cmds:
|
||||||
|
- echo "Building {{.OUTPUT_FILE}}..."
|
||||||
|
- go build {{.GO_FLAGS}} -ldflags="{{.LDFLAGS}}" -o "{{.OUTPUT_FILE}}" {{.MAIN_FILE}}
|
||||||
|
|
||||||
|
# Install task
|
||||||
|
install:
|
||||||
|
desc: Install the binary to $GOPATH/bin
|
||||||
|
deps: [test]
|
||||||
|
cmds:
|
||||||
|
- go install -ldflags="{{.LDFLAGS}}" {{.MAIN_FILE}}
|
||||||
|
- echo "Installed {{.APP_NAME}} to $(go env GOPATH)/bin"
|
||||||
|
|
||||||
|
# Testing tasks
|
||||||
|
test:
|
||||||
|
desc: Run all tests
|
||||||
|
aliases: [t]
|
||||||
|
env:
|
||||||
|
CGO_ENABLED: 1
|
||||||
|
cmds:
|
||||||
|
- go test {{.GO_FLAGS}} -race -timeout {{.TEST_TIMEOUT}} ./...
|
||||||
|
|
||||||
|
test:coverage:
|
||||||
|
desc: Run tests with coverage report
|
||||||
|
aliases: [cover, cov]
|
||||||
|
deps: [clean-coverage]
|
||||||
|
env:
|
||||||
|
CGO_ENABLED: 1
|
||||||
|
cmds:
|
||||||
|
- task: mkdir
|
||||||
|
vars: { DIR: "{{.COVERAGE_DIR}}" }
|
||||||
|
- go test {{.GO_FLAGS}} -race -coverprofile={{.COVERAGE_DIR}}/coverage.out -covermode=atomic -timeout {{.TEST_TIMEOUT}} ./...
|
||||||
|
- go tool cover -html={{.COVERAGE_DIR}}/coverage.out -o {{.COVERAGE_DIR}}/coverage.html
|
||||||
|
- go tool cover -func={{.COVERAGE_DIR}}/coverage.out
|
||||||
|
- echo "Coverage report generated at {{.COVERAGE_DIR}}/coverage.html"
|
||||||
|
|
||||||
|
test:verbose:
|
||||||
|
desc: Run tests with verbose output
|
||||||
|
aliases: [tv]
|
||||||
|
env:
|
||||||
|
CGO_ENABLED: 1
|
||||||
|
cmds:
|
||||||
|
- go test -v -race -timeout {{.TEST_TIMEOUT}} ./...
|
||||||
|
|
||||||
|
test:watch:
|
||||||
|
desc: Run tests in watch mode
|
||||||
|
aliases: [tw]
|
||||||
|
watch: true
|
||||||
|
sources:
|
||||||
|
- "**/*.go"
|
||||||
|
cmds:
|
||||||
|
- task: test
|
||||||
|
|
||||||
|
test:bench:
|
||||||
|
desc: Run benchmark tests
|
||||||
|
aliases: [bench]
|
||||||
|
cmds:
|
||||||
|
- go test -bench=. -benchmem -timeout {{.TEST_TIMEOUT}} ./...
|
||||||
|
|
||||||
|
test:integration:
|
||||||
|
desc: Run integration tests
|
||||||
|
env:
|
||||||
|
CGO_ENABLED: 1
|
||||||
|
status:
|
||||||
|
- '{{if eq OS "windows"}}if not exist "main_test.go" exit 1{{else}}test ! -f "main_test.go"{{end}}'
|
||||||
|
cmds:
|
||||||
|
- go test -v -race -tags=integration -timeout {{.TEST_TIMEOUT}} ./...
|
||||||
|
|
||||||
|
# Code quality tasks
|
||||||
|
lint:
|
||||||
|
desc: Run all linters
|
||||||
|
silent: true
|
||||||
|
aliases: [l]
|
||||||
|
cmds:
|
||||||
|
- task: lint:vet
|
||||||
|
- task: lint:fmt
|
||||||
|
- task: lint:staticcheck
|
||||||
|
- task: lint:golangci
|
||||||
|
|
||||||
|
lint:vet:
|
||||||
|
desc: Run go vet
|
||||||
|
silent: true
|
||||||
|
cmds:
|
||||||
|
- go vet ./...
|
||||||
|
|
||||||
|
lint:fmt:
|
||||||
|
desc: Check code formatting
|
||||||
|
silent: true
|
||||||
|
vars:
|
||||||
|
UNFORMATTED:
|
||||||
|
sh: gofmt -s -l .
|
||||||
|
cmds:
|
||||||
|
- |
|
||||||
|
{{if ne .UNFORMATTED ""}}
|
||||||
|
echo "❌ The following files need formatting:"
|
||||||
|
echo "{{.UNFORMATTED}}"
|
||||||
|
exit 1
|
||||||
|
{{else}}
|
||||||
|
echo "All files are properly formatted"
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
lint:staticcheck:
|
||||||
|
desc: Run staticcheck (install if needed)
|
||||||
|
silent: true
|
||||||
|
vars:
|
||||||
|
HAS_STATICCHECK:
|
||||||
|
sh: '{{if eq OS "windows"}}where.exe staticcheck 2>NUL{{else}}command -v staticcheck 2>/dev/null{{end}}'
|
||||||
|
cmds:
|
||||||
|
- '{{if eq .HAS_STATICCHECK ""}}echo "Installing staticcheck..." && go install honnef.co/go/tools/cmd/staticcheck@latest{{end}}'
|
||||||
|
- staticcheck ./...
|
||||||
|
ignore_error: true
|
||||||
|
|
||||||
|
lint:golangci:
|
||||||
|
desc: Run golangci-lint (install if needed)
|
||||||
|
silent: true
|
||||||
|
aliases: [golangci, golangci-lint]
|
||||||
|
vars:
|
||||||
|
HAS_GOLANGCI:
|
||||||
|
sh: '{{if eq OS "windows"}}where.exe golangci-lint 2>NUL{{else}}command -v golangci-lint 2>/dev/null{{end}}'
|
||||||
|
cmds:
|
||||||
|
- '{{if eq .HAS_GOLANGCI ""}}echo "Installing golangci-lint..." && go install github.com/golangci/golangci-lint/cmd/golangci-lint@latest{{end}}'
|
||||||
|
- golangci-lint run ./...
|
||||||
|
- echo "✅ golangci-lint passed"
|
||||||
|
|
||||||
|
lint:golangci:fix:
|
||||||
|
desc: Run golangci-lint with auto-fix
|
||||||
|
silent: true
|
||||||
|
aliases: [golangci-fix]
|
||||||
|
vars:
|
||||||
|
HAS_GOLANGCI:
|
||||||
|
sh: '{{if eq OS "windows"}}where.exe golangci-lint 2>NUL{{else}}command -v golangci-lint 2>/dev/null{{end}}'
|
||||||
|
cmds:
|
||||||
|
- '{{if eq .HAS_GOLANGCI ""}}echo "Installing golangci-lint..." && go install github.com/golangci/golangci-lint/cmd/golangci-lint@latest{{end}}'
|
||||||
|
- golangci-lint run --fix ./...
|
||||||
|
- echo "golangci-lint fixes applied"
|
||||||
|
|
||||||
|
fmt:
|
||||||
|
desc: Format all Go files
|
||||||
|
silent: true
|
||||||
|
aliases: [format]
|
||||||
|
cmds:
|
||||||
|
- gofmt -s -w .
|
||||||
|
- echo "Formatted all Go files"
|
||||||
|
|
||||||
|
modernize:
|
||||||
|
desc: Modernize Go code to use modern idioms
|
||||||
|
silent: true
|
||||||
|
aliases: [modern]
|
||||||
|
cmds:
|
||||||
|
- go run golang.org/x/tools/gopls/internal/analysis/modernize/cmd/modernize@latest -fix -test ./...
|
||||||
|
- echo "Code modernized"
|
||||||
|
|
||||||
|
# Dependency management
|
||||||
|
deps:
|
||||||
|
desc: Download and verify dependencies
|
||||||
|
aliases: [mod]
|
||||||
|
cmds:
|
||||||
|
- go mod download
|
||||||
|
- go mod verify
|
||||||
|
- echo "Dependencies downloaded and verified"
|
||||||
|
|
||||||
|
deps:tidy:
|
||||||
|
desc: Tidy go.mod and go.sum
|
||||||
|
aliases: [tidy]
|
||||||
|
cmds:
|
||||||
|
- go mod tidy
|
||||||
|
- echo "Dependencies tidied"
|
||||||
|
|
||||||
|
deps:update:
|
||||||
|
desc: Update all dependencies to latest versions
|
||||||
|
aliases: [update]
|
||||||
|
cmds:
|
||||||
|
- go get -u ./...
|
||||||
|
- go mod tidy
|
||||||
|
- echo "Dependencies updated"
|
||||||
|
|
||||||
|
deps:graph:
|
||||||
|
desc: Display dependency graph
|
||||||
|
cmds:
|
||||||
|
- go mod graph
|
||||||
|
|
||||||
|
# Docker tasks
|
||||||
|
docker:build:
|
||||||
|
desc: Build Docker image
|
||||||
|
aliases: [db]
|
||||||
|
cmds:
|
||||||
|
- |
|
||||||
|
docker build \
|
||||||
|
--build-arg VERSION={{.VERSION}} \
|
||||||
|
--build-arg BUILD_TIME={{.BUILD_TIME}} \
|
||||||
|
--build-arg GIT_COMMIT={{.GIT_COMMIT}} \
|
||||||
|
-t {{.APP_NAME}}:{{.VERSION}} \
|
||||||
|
-t {{.APP_NAME}}:latest \
|
||||||
|
.
|
||||||
|
- >
|
||||||
|
echo "Docker image built: {{.APP_NAME}}:{{.VERSION}}"
|
||||||
|
|
||||||
|
docker:build:dev:
|
||||||
|
desc: Build development Docker image
|
||||||
|
cmds:
|
||||||
|
- docker build -f Dockerfile.dev -t {{.APP_NAME}}:dev .
|
||||||
|
- >
|
||||||
|
echo "Development Docker image built: {{.APP_NAME}}:dev"
|
||||||
|
|
||||||
|
docker:run:
|
||||||
|
desc: Run Docker container
|
||||||
|
aliases: [dr]
|
||||||
|
deps: [docker:build]
|
||||||
|
cmds:
|
||||||
|
- docker run --rm {{.APP_NAME}}:{{.VERSION}} --help
|
||||||
|
|
||||||
|
docker:test:
|
||||||
|
desc: Test Docker image
|
||||||
|
deps: [docker:build]
|
||||||
|
cmds:
|
||||||
|
- docker run --rm {{.APP_NAME}}:{{.VERSION}} --version
|
||||||
|
- echo "Docker image tested successfully"
|
||||||
|
|
||||||
|
docker:compose:up:
|
||||||
|
desc: Start services with docker-compose
|
||||||
|
cmds:
|
||||||
|
- docker-compose up -d
|
||||||
|
|
||||||
|
docker:compose:down:
|
||||||
|
desc: Stop services with docker-compose
|
||||||
|
cmds:
|
||||||
|
- docker-compose down
|
||||||
|
|
||||||
|
# Cleanup tasks
|
||||||
|
clean:
|
||||||
|
desc: Clean all generated files
|
||||||
|
aliases: [c]
|
||||||
|
cmds:
|
||||||
|
- task: clean-bin
|
||||||
|
- task: clean-coverage
|
||||||
|
- task: clean-cache
|
||||||
|
- echo "All generated files cleaned"
|
||||||
|
|
||||||
|
clean-bin:
|
||||||
|
desc: Remove built binaries
|
||||||
|
internal: true
|
||||||
|
cmds:
|
||||||
|
- task: rmdir
|
||||||
|
vars: { DIR: "{{.OUTPUT_DIR}}" }
|
||||||
|
|
||||||
|
clean-coverage:
|
||||||
|
desc: Remove coverage files
|
||||||
|
internal: true
|
||||||
|
cmds:
|
||||||
|
- task: rmdir
|
||||||
|
vars: { DIR: "{{.COVERAGE_DIR}}" }
|
||||||
|
|
||||||
|
clean-cache:
|
||||||
|
desc: Clean Go build and test cache
|
||||||
|
cmds:
|
||||||
|
- go clean -cache -testcache -modcache
|
||||||
|
- echo "Go caches cleaned"
|
||||||
|
|
||||||
|
# CI/CD tasks
|
||||||
|
ci:
|
||||||
|
desc: Run all CI checks (test, lint, build)
|
||||||
|
cmds:
|
||||||
|
- task: deps
|
||||||
|
- task: lint
|
||||||
|
- task: test:coverage
|
||||||
|
- task: build
|
||||||
|
- echo "All CI checks passed"
|
||||||
|
|
||||||
|
ci:local:
|
||||||
|
desc: Run CI checks locally with detailed output
|
||||||
|
cmds:
|
||||||
|
- echo "🔍 Running local CI checks..."
|
||||||
|
- echo ""
|
||||||
|
- echo "📦 Checking dependencies..."
|
||||||
|
- task: deps
|
||||||
|
- echo ""
|
||||||
|
- echo "🔧 Running linters..."
|
||||||
|
- task: lint
|
||||||
|
- echo ""
|
||||||
|
- echo "🧪 Running tests with coverage..."
|
||||||
|
- task: test:coverage
|
||||||
|
- echo ""
|
||||||
|
- echo "🏗️ Building application..."
|
||||||
|
- task: build:all
|
||||||
|
- echo ""
|
||||||
|
- echo "All CI checks completed successfully!"
|
||||||
|
|
||||||
|
# Release tasks
|
||||||
|
release:check:
|
||||||
|
desc: Check if ready for release
|
||||||
|
cmds:
|
||||||
|
- task: ci
|
||||||
|
- git diff --exit-code
|
||||||
|
- git diff --cached --exit-code
|
||||||
|
- echo "Ready for release"
|
||||||
|
|
||||||
|
release:tag:
|
||||||
|
desc: Tag a new release (requires VERSION env var)
|
||||||
|
requires:
|
||||||
|
vars: [VERSION]
|
||||||
|
preconditions:
|
||||||
|
- sh: "git diff --exit-code"
|
||||||
|
msg: "Working directory is not clean"
|
||||||
|
- sh: "git diff --cached --exit-code"
|
||||||
|
msg: "Staging area is not clean"
|
||||||
|
cmds:
|
||||||
|
- git tag -a v{{.VERSION}} -m "Release v{{.VERSION}}"
|
||||||
|
- echo "Tagged v{{.VERSION}}"
|
||||||
|
- >
|
||||||
|
echo "Push with: git push origin v{{.VERSION}}"
|
||||||
|
|
||||||
|
# Documentation tasks
|
||||||
|
docs:serve:
|
||||||
|
desc: Serve documentation locally
|
||||||
|
vars:
|
||||||
|
HAS_GODOC:
|
||||||
|
sh: '{{if eq OS "windows"}}where.exe godoc 2>NUL{{else}}command -v godoc 2>/dev/null{{end}}'
|
||||||
|
cmds:
|
||||||
|
- '{{if eq .HAS_GODOC ""}}echo "Installing godoc..." && go install golang.org/x/tools/cmd/godoc@latest{{end}}'
|
||||||
|
- echo "📚 Serving documentation at http://localhost:6060"
|
||||||
|
- godoc -http=:6060
|
||||||
|
interactive: true
|
||||||
|
|
||||||
|
docs:coverage:
|
||||||
|
desc: Open coverage report in browser
|
||||||
|
deps: [test:coverage]
|
||||||
|
cmds:
|
||||||
|
- '{{if eq OS "darwin"}}open {{.COVERAGE_DIR}}/coverage.html{{else if eq OS "windows"}}start {{.COVERAGE_DIR}}/coverage.html{{else}}xdg-open {{.COVERAGE_DIR}}/coverage.html 2>/dev/null || echo "Please open {{.COVERAGE_DIR}}/coverage.html in your browser"{{end}}'
|
||||||
|
|
||||||
|
# Info tasks
|
||||||
|
info:
|
||||||
|
desc: Display build information
|
||||||
|
vars:
|
||||||
|
GO_VERSION:
|
||||||
|
sh: go version
|
||||||
|
cmds:
|
||||||
|
- task: info:print
|
||||||
|
silent: true
|
||||||
|
|
||||||
|
info:print:
|
||||||
|
internal: true
|
||||||
|
silent: true
|
||||||
|
vars:
|
||||||
|
GO_VERSION:
|
||||||
|
sh: go version
|
||||||
|
cmds:
|
||||||
|
- echo "Application Info:"
|
||||||
|
- echo " Name{{":"}} {{.APP_NAME}}"
|
||||||
|
- echo " Version{{":"}} {{.VERSION}}"
|
||||||
|
- echo " Git Commit{{":"}} {{.GIT_COMMIT}}"
|
||||||
|
- echo " Build Time{{":"}} {{.BUILD_TIME}}"
|
||||||
|
- echo ""
|
||||||
|
- echo "Go Environment{{":"}}"
|
||||||
|
- echo " Go Version{{":"}} {{.GO_VERSION}}"
|
||||||
|
- echo " GOOS{{":"}} {{.GOOS}}"
|
||||||
|
- echo " GOARCH{{":"}} {{.GOARCH}}"
|
||||||
|
- echo " CGO{{":"}} {{.CGO_ENABLED}}"
|
||||||
|
- echo ""
|
||||||
|
- echo "Paths{{":"}}"
|
||||||
|
- echo " Output Dir{{":"}} {{.OUTPUT_DIR}}"
|
||||||
|
- echo " Coverage{{":"}} {{.COVERAGE_DIR}}"
|
||||||
|
|
||||||
|
# Security tasks
|
||||||
|
security:check:
|
||||||
|
desc: Run security checks with gosec
|
||||||
|
vars:
|
||||||
|
HAS_GOSEC:
|
||||||
|
sh: '{{if eq OS "windows"}}where.exe gosec 2>NUL{{else}}command -v gosec 2>/dev/null{{end}}'
|
||||||
|
cmds:
|
||||||
|
- '{{if eq .HAS_GOSEC ""}}echo "Installing gosec..." && go install github.com/securego/gosec/v2/cmd/gosec@latest{{end}}'
|
||||||
|
- gosec ./...
|
||||||
|
ignore_error: true
|
||||||
|
|
||||||
|
security:audit:
|
||||||
|
desc: Audit dependencies for vulnerabilities
|
||||||
|
vars:
|
||||||
|
HAS_GOVULNCHECK:
|
||||||
|
sh: '{{if eq OS "windows"}}where.exe govulncheck 2>NUL{{else}}command -v govulncheck 2>/dev/null{{end}}'
|
||||||
|
cmds:
|
||||||
|
- '{{if eq .HAS_GOVULNCHECK ""}}echo "Installing govulncheck..." && go install golang.org/x/vuln/cmd/govulncheck@latest{{end}}'
|
||||||
|
- govulncheck ./...
|
||||||
|
|
||||||
|
# Example/Demo tasks
|
||||||
|
demo:markdown:
|
||||||
|
desc: Demo - Convert sample to Markdown
|
||||||
|
status:
|
||||||
|
- '{{if eq OS "windows"}}if not exist "articulate-sample.json" exit 1{{else}}test ! -f "articulate-sample.json"{{end}}'
|
||||||
|
deps: [build]
|
||||||
|
cmds:
|
||||||
|
- "{{.OUTPUT_DIR}}/{{.APP_NAME}}{{.EXE_EXT}} articulate-sample.json md output-demo.md"
|
||||||
|
- echo "Demo Markdown created{{:}} output-demo.md"
|
||||||
|
- defer:
|
||||||
|
task: rmfile
|
||||||
|
vars: { FILE: "output-demo.md" }
|
||||||
|
|
||||||
|
demo:html:
|
||||||
|
desc: Demo - Convert sample to HTML
|
||||||
|
status:
|
||||||
|
- '{{if eq OS "windows"}}if not exist "articulate-sample.json" exit 1{{else}}test ! -f "articulate-sample.json"{{end}}'
|
||||||
|
deps: [build]
|
||||||
|
cmds:
|
||||||
|
- "{{.OUTPUT_DIR}}/{{.APP_NAME}}{{.EXE_EXT}} articulate-sample.json html output-demo.html"
|
||||||
|
- echo "Demo HTML created{{:}} output-demo.html"
|
||||||
|
- defer:
|
||||||
|
task: rmfile
|
||||||
|
vars: { FILE: "output-demo.html" }
|
||||||
|
|
||||||
|
demo:docx:
|
||||||
|
desc: Demo - Convert sample to DOCX
|
||||||
|
status:
|
||||||
|
- '{{if eq OS "windows"}}if not exist "articulate-sample.json" exit 1{{else}}test ! -f "articulate-sample.json"{{end}}'
|
||||||
|
deps: [build]
|
||||||
|
cmds:
|
||||||
|
- "{{.OUTPUT_DIR}}/{{.APP_NAME}}{{.EXE_EXT}} articulate-sample.json docx output-demo.docx"
|
||||||
|
- echo "Demo DOCX created{{:}} output-demo.docx"
|
||||||
|
- defer:
|
||||||
|
task: rmfile
|
||||||
|
vars: { FILE: "output-demo.docx" }
|
||||||
|
|
||||||
|
# Performance profiling
|
||||||
|
profile:cpu:
|
||||||
|
desc: Run CPU profiling
|
||||||
|
cmds:
|
||||||
|
- go test -cpuprofile=cpu.prof -bench=. ./...
|
||||||
|
- go tool pprof -http=:8080 cpu.prof
|
||||||
|
- defer:
|
||||||
|
task: rmfile
|
||||||
|
vars: { FILE: "cpu.prof" }
|
||||||
|
|
||||||
|
profile:mem:
|
||||||
|
desc: Run memory profiling
|
||||||
|
cmds:
|
||||||
|
- go test -memprofile=mem.prof -bench=. ./...
|
||||||
|
- go tool pprof -http=:8080 mem.prof
|
||||||
|
- defer:
|
||||||
|
task: rmfile
|
||||||
|
vars: { FILE: "mem.prof" }
|
||||||
|
|
||||||
|
# Git hooks
|
||||||
|
hooks:install:
|
||||||
|
desc: Install git hooks
|
||||||
|
cmds:
|
||||||
|
- task: mkdir
|
||||||
|
vars: { DIR: ".git/hooks" }
|
||||||
|
- '{{if eq OS "windows"}}echo "#!/bin/sh" > .git/hooks/pre-commit && echo "task lint:fmt" >> .git/hooks/pre-commit{{else}}cat > .git/hooks/pre-commit << ''EOF''{{printf "\n"}}#!/bin/sh{{printf "\n"}}task lint:fmt{{printf "\n"}}EOF{{printf "\n"}}chmod +x .git/hooks/pre-commit{{end}}'
|
||||||
|
- echo "Git hooks installed"
|
||||||
|
|
||||||
|
# Quick shortcuts
|
||||||
|
qa:
|
||||||
|
desc: Quick quality assurance (fmt + lint + test)
|
||||||
|
aliases: [q, quick]
|
||||||
|
cmds:
|
||||||
|
- task: fmt
|
||||||
|
- task: lint
|
||||||
|
- task: test
|
||||||
|
- echo "Quick QA passed"
|
||||||
|
|
||||||
|
all:
|
||||||
|
desc: Build everything (clean + deps + test + build:all + docker:build)
|
||||||
|
cmds:
|
||||||
|
- task: clean
|
||||||
|
- task: deps:tidy
|
||||||
|
- task: test:coverage
|
||||||
|
- task: build:all
|
||||||
|
- task: docker:build
|
||||||
|
- echo "Full build completed!"
|
||||||
|
|
||||||
|
# Cross-platform helper tasks
|
||||||
|
mkdir:
|
||||||
|
internal: true
|
||||||
|
requires:
|
||||||
|
vars: [DIR]
|
||||||
|
cmds:
|
||||||
|
- '{{if eq OS "windows"}}powershell -Command "New-Item -ItemType Directory -Force -Path ''{{.DIR}}'' | Out-Null"{{else}}mkdir -p "{{.DIR}}"{{end}}'
|
||||||
|
silent: true
|
||||||
|
|
||||||
|
rmdir:
|
||||||
|
internal: true
|
||||||
|
requires:
|
||||||
|
vars: [DIR]
|
||||||
|
cmds:
|
||||||
|
- '{{if eq OS "windows"}}powershell -Command "if (Test-Path ''{{.DIR}}'') { Remove-Item -Recurse -Force ''{{.DIR}}'' }"{{else}}rm -rf "{{.DIR}}" 2>/dev/null || true{{end}}'
|
||||||
|
silent: true
|
||||||
|
|
||||||
|
rmfile:
|
||||||
|
internal: true
|
||||||
|
requires:
|
||||||
|
vars: [FILE]
|
||||||
|
cmds:
|
||||||
|
- '{{if eq OS "windows"}}powershell -Command "if (Test-Path ''{{.FILE}}'') { Remove-Item -Force ''{{.FILE}}'' }"{{else}}rm -f "{{.FILE}}"{{end}}'
|
||||||
|
silent: true
|
||||||
12
go.mod
12
go.mod
@ -1,10 +1,16 @@
|
|||||||
module github.com/kjanat/articulate-parser
|
module github.com/kjanat/articulate-parser
|
||||||
|
|
||||||
go 1.23.0
|
go 1.24.0
|
||||||
|
|
||||||
require github.com/fumiama/go-docx v0.0.0-20250506085032-0c30fd09304b
|
toolchain go1.25.5
|
||||||
|
|
||||||
|
require (
|
||||||
|
github.com/fumiama/go-docx v0.0.0-20250506085032-0c30fd09304b
|
||||||
|
golang.org/x/net v0.48.0
|
||||||
|
golang.org/x/text v0.33.0
|
||||||
|
)
|
||||||
|
|
||||||
require (
|
require (
|
||||||
github.com/fumiama/imgsz v0.0.4 // indirect
|
github.com/fumiama/imgsz v0.0.4 // indirect
|
||||||
golang.org/x/image v0.27.0 // indirect
|
golang.org/x/image v0.34.0 // indirect
|
||||||
)
|
)
|
||||||
|
|||||||
8
go.sum
8
go.sum
@ -2,5 +2,9 @@ github.com/fumiama/go-docx v0.0.0-20250506085032-0c30fd09304b h1:/mxSugRc4SgN7Xg
|
|||||||
github.com/fumiama/go-docx v0.0.0-20250506085032-0c30fd09304b/go.mod h1:ssRF0IaB1hCcKIObp3FkZOsjTcAHpgii70JelNb4H8M=
|
github.com/fumiama/go-docx v0.0.0-20250506085032-0c30fd09304b/go.mod h1:ssRF0IaB1hCcKIObp3FkZOsjTcAHpgii70JelNb4H8M=
|
||||||
github.com/fumiama/imgsz v0.0.4 h1:Lsasu2hdSSFS+vnD+nvR1UkiRMK7hcpyYCC0FzgSMFI=
|
github.com/fumiama/imgsz v0.0.4 h1:Lsasu2hdSSFS+vnD+nvR1UkiRMK7hcpyYCC0FzgSMFI=
|
||||||
github.com/fumiama/imgsz v0.0.4/go.mod h1:bISOQVTlw9sRytPwe8ir7tAaEmyz9hSNj9n8mXMBG0E=
|
github.com/fumiama/imgsz v0.0.4/go.mod h1:bISOQVTlw9sRytPwe8ir7tAaEmyz9hSNj9n8mXMBG0E=
|
||||||
golang.org/x/image v0.27.0 h1:C8gA4oWU/tKkdCfYT6T2u4faJu3MeNS5O8UPWlPF61w=
|
golang.org/x/image v0.34.0 h1:33gCkyw9hmwbZJeZkct8XyR11yH889EQt/QH4VmXMn8=
|
||||||
golang.org/x/image v0.27.0/go.mod h1:xbdrClrAUway1MUTEZDq9mz/UpRwYAkFFNUslZtcB+g=
|
golang.org/x/image v0.34.0/go.mod h1:2RNFBZRB+vnwwFil8GkMdRvrJOFd1AzdZI6vOY+eJVU=
|
||||||
|
golang.org/x/net v0.48.0 h1:zyQRTTrjc33Lhh0fBgT/H3oZq9WuvRR5gPC70xpDiQU=
|
||||||
|
golang.org/x/net v0.48.0/go.mod h1:+ndRgGjkh8FGtu1w1FGbEC31if4VrNVMuKTgcAAnQRY=
|
||||||
|
golang.org/x/text v0.33.0 h1:B3njUFyqtHDUI5jMn1YIr5B0IE2U0qck04r6d4KPAxE=
|
||||||
|
golang.org/x/text v0.33.0/go.mod h1:LuMebE6+rBincTi9+xWTY8TztLzKHc/9C1uBCG27+q8=
|
||||||
|
|||||||
77
internal/config/config.go
Normal file
77
internal/config/config.go
Normal file
@ -0,0 +1,77 @@
|
|||||||
|
// Package config provides configuration management for the articulate-parser application.
|
||||||
|
// It supports loading configuration from environment variables and command-line flags.
|
||||||
|
package config
|
||||||
|
|
||||||
|
import (
|
||||||
|
"log/slog"
|
||||||
|
"os"
|
||||||
|
"strconv"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Config holds all configuration values for the application.
|
||||||
|
type Config struct {
|
||||||
|
// Parser configuration
|
||||||
|
BaseURL string
|
||||||
|
RequestTimeout time.Duration
|
||||||
|
|
||||||
|
// Logging configuration
|
||||||
|
LogLevel slog.Level
|
||||||
|
LogFormat string // "json" or "text"
|
||||||
|
}
|
||||||
|
|
||||||
|
// Default configuration values.
|
||||||
|
const (
|
||||||
|
DefaultBaseURL = "https://rise.articulate.com"
|
||||||
|
DefaultRequestTimeout = 30 * time.Second
|
||||||
|
DefaultLogLevel = slog.LevelInfo
|
||||||
|
DefaultLogFormat = "text"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Load creates a new Config with values from environment variables.
|
||||||
|
// Falls back to defaults if environment variables are not set.
|
||||||
|
func Load() *Config {
|
||||||
|
return &Config{
|
||||||
|
BaseURL: getEnv("ARTICULATE_BASE_URL", DefaultBaseURL),
|
||||||
|
RequestTimeout: getDurationEnv("ARTICULATE_REQUEST_TIMEOUT", DefaultRequestTimeout),
|
||||||
|
LogLevel: getLogLevelEnv("LOG_LEVEL", DefaultLogLevel),
|
||||||
|
LogFormat: getEnv("LOG_FORMAT", DefaultLogFormat),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// getEnv retrieves an environment variable or returns the default value.
|
||||||
|
func getEnv(key, defaultValue string) string {
|
||||||
|
if value := os.Getenv(key); value != "" {
|
||||||
|
return value
|
||||||
|
}
|
||||||
|
return defaultValue
|
||||||
|
}
|
||||||
|
|
||||||
|
// getDurationEnv retrieves a duration from environment variable or returns default.
|
||||||
|
// The environment variable should be in seconds (e.g., "30" for 30 seconds).
|
||||||
|
func getDurationEnv(key string, defaultValue time.Duration) time.Duration {
|
||||||
|
if value := os.Getenv(key); value != "" {
|
||||||
|
if seconds, err := strconv.Atoi(value); err == nil {
|
||||||
|
return time.Duration(seconds) * time.Second
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return defaultValue
|
||||||
|
}
|
||||||
|
|
||||||
|
// getLogLevelEnv retrieves a log level from environment variable or returns default.
|
||||||
|
// Accepts: "debug", "info", "warn", "error" (case-insensitive).
|
||||||
|
func getLogLevelEnv(key string, defaultValue slog.Level) slog.Level {
|
||||||
|
value := os.Getenv(key)
|
||||||
|
switch value {
|
||||||
|
case "debug", "DEBUG":
|
||||||
|
return slog.LevelDebug
|
||||||
|
case "info", "INFO":
|
||||||
|
return slog.LevelInfo
|
||||||
|
case "warn", "WARN", "warning", "WARNING":
|
||||||
|
return slog.LevelWarn
|
||||||
|
case "error", "ERROR":
|
||||||
|
return slog.LevelError
|
||||||
|
default:
|
||||||
|
return defaultValue
|
||||||
|
}
|
||||||
|
}
|
||||||
116
internal/config/config_test.go
Normal file
116
internal/config/config_test.go
Normal file
@ -0,0 +1,116 @@
|
|||||||
|
package config
|
||||||
|
|
||||||
|
import (
|
||||||
|
"log/slog"
|
||||||
|
"os"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestLoad(t *testing.T) {
|
||||||
|
// Clear environment
|
||||||
|
os.Clearenv()
|
||||||
|
|
||||||
|
cfg := Load()
|
||||||
|
|
||||||
|
if cfg.BaseURL != DefaultBaseURL {
|
||||||
|
t.Errorf("Expected BaseURL '%s', got '%s'", DefaultBaseURL, cfg.BaseURL)
|
||||||
|
}
|
||||||
|
|
||||||
|
if cfg.RequestTimeout != DefaultRequestTimeout {
|
||||||
|
t.Errorf("Expected timeout %v, got %v", DefaultRequestTimeout, cfg.RequestTimeout)
|
||||||
|
}
|
||||||
|
|
||||||
|
if cfg.LogLevel != DefaultLogLevel {
|
||||||
|
t.Errorf("Expected log level %v, got %v", DefaultLogLevel, cfg.LogLevel)
|
||||||
|
}
|
||||||
|
|
||||||
|
if cfg.LogFormat != DefaultLogFormat {
|
||||||
|
t.Errorf("Expected log format '%s', got '%s'", DefaultLogFormat, cfg.LogFormat)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestLoad_WithEnvironmentVariables(t *testing.T) {
|
||||||
|
// Set environment variables
|
||||||
|
t.Setenv("ARTICULATE_BASE_URL", "https://test.example.com")
|
||||||
|
t.Setenv("ARTICULATE_REQUEST_TIMEOUT", "60")
|
||||||
|
t.Setenv("LOG_LEVEL", "debug")
|
||||||
|
t.Setenv("LOG_FORMAT", "json")
|
||||||
|
|
||||||
|
cfg := Load()
|
||||||
|
|
||||||
|
if cfg.BaseURL != "https://test.example.com" {
|
||||||
|
t.Errorf("Expected BaseURL 'https://test.example.com', got '%s'", cfg.BaseURL)
|
||||||
|
}
|
||||||
|
|
||||||
|
if cfg.RequestTimeout != 60*time.Second {
|
||||||
|
t.Errorf("Expected timeout 60s, got %v", cfg.RequestTimeout)
|
||||||
|
}
|
||||||
|
|
||||||
|
if cfg.LogLevel != slog.LevelDebug {
|
||||||
|
t.Errorf("Expected log level Debug, got %v", cfg.LogLevel)
|
||||||
|
}
|
||||||
|
|
||||||
|
if cfg.LogFormat != "json" {
|
||||||
|
t.Errorf("Expected log format 'json', got '%s'", cfg.LogFormat)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestGetLogLevelEnv(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
value string
|
||||||
|
expected slog.Level
|
||||||
|
}{
|
||||||
|
{"debug lowercase", "debug", slog.LevelDebug},
|
||||||
|
{"debug uppercase", "DEBUG", slog.LevelDebug},
|
||||||
|
{"info lowercase", "info", slog.LevelInfo},
|
||||||
|
{"info uppercase", "INFO", slog.LevelInfo},
|
||||||
|
{"warn lowercase", "warn", slog.LevelWarn},
|
||||||
|
{"warn uppercase", "WARN", slog.LevelWarn},
|
||||||
|
{"warning lowercase", "warning", slog.LevelWarn},
|
||||||
|
{"error lowercase", "error", slog.LevelError},
|
||||||
|
{"error uppercase", "ERROR", slog.LevelError},
|
||||||
|
{"invalid value", "invalid", slog.LevelInfo},
|
||||||
|
{"empty value", "", slog.LevelInfo},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
os.Clearenv()
|
||||||
|
if tt.value != "" {
|
||||||
|
t.Setenv("TEST_LOG_LEVEL", tt.value)
|
||||||
|
}
|
||||||
|
result := getLogLevelEnv("TEST_LOG_LEVEL", slog.LevelInfo)
|
||||||
|
if result != tt.expected {
|
||||||
|
t.Errorf("Expected %v, got %v", tt.expected, result)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestGetDurationEnv(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
value string
|
||||||
|
expected time.Duration
|
||||||
|
}{
|
||||||
|
{"valid duration", "45", 45 * time.Second},
|
||||||
|
{"zero duration", "0", 0},
|
||||||
|
{"invalid duration", "invalid", 30 * time.Second},
|
||||||
|
{"empty value", "", 30 * time.Second},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
os.Clearenv()
|
||||||
|
if tt.value != "" {
|
||||||
|
t.Setenv("TEST_DURATION", tt.value)
|
||||||
|
}
|
||||||
|
result := getDurationEnv("TEST_DURATION", 30*time.Second)
|
||||||
|
if result != tt.expected {
|
||||||
|
t.Errorf("Expected %v, got %v", tt.expected, result)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
200
internal/exporters/bench_test.go
Normal file
200
internal/exporters/bench_test.go
Normal file
@ -0,0 +1,200 @@
|
|||||||
|
package exporters
|
||||||
|
|
||||||
|
import (
|
||||||
|
"path/filepath"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
|
)
|
||||||
|
|
||||||
|
// BenchmarkFactory_CreateExporter_Markdown benchmarks markdown exporter creation.
|
||||||
|
func BenchmarkFactory_CreateExporter_Markdown(b *testing.B) {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
factory := NewFactory(htmlCleaner)
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
_, _ = factory.CreateExporter("markdown")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkFactory_CreateExporter_All benchmarks creating all exporter types.
|
||||||
|
func BenchmarkFactory_CreateExporter_All(b *testing.B) {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
factory := NewFactory(htmlCleaner)
|
||||||
|
formats := []string{"markdown", "docx", "html"}
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
for _, format := range formats {
|
||||||
|
_, _ = factory.CreateExporter(format)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkAllExporters_Export benchmarks all exporters with the same course.
|
||||||
|
func BenchmarkAllExporters_Export(b *testing.B) {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
course := createBenchmarkCourse()
|
||||||
|
|
||||||
|
exporters := map[string]struct {
|
||||||
|
exporter any
|
||||||
|
ext string
|
||||||
|
}{
|
||||||
|
"Markdown": {NewMarkdownExporter(htmlCleaner), ".md"},
|
||||||
|
"Docx": {NewDocxExporter(htmlCleaner), ".docx"},
|
||||||
|
"HTML": {NewHTMLExporter(htmlCleaner), ".html"},
|
||||||
|
}
|
||||||
|
|
||||||
|
for name, exp := range exporters {
|
||||||
|
b.Run(name, func(b *testing.B) {
|
||||||
|
tempDir := b.TempDir()
|
||||||
|
exporter := exp.exporter.(interface {
|
||||||
|
Export(*models.Course, string) error
|
||||||
|
})
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
outputPath := filepath.Join(tempDir, "benchmark"+exp.ext)
|
||||||
|
_ = exporter.Export(course, outputPath)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkExporters_LargeCourse benchmarks exporters with large course data.
|
||||||
|
func BenchmarkExporters_LargeCourse(b *testing.B) {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
course := createLargeBenchmarkCourse()
|
||||||
|
|
||||||
|
b.Run("Markdown_Large", func(b *testing.B) {
|
||||||
|
exporter := NewMarkdownExporter(htmlCleaner)
|
||||||
|
tempDir := b.TempDir()
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
outputPath := filepath.Join(tempDir, "large.md")
|
||||||
|
_ = exporter.Export(course, outputPath)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
b.Run("Docx_Large", func(b *testing.B) {
|
||||||
|
exporter := NewDocxExporter(htmlCleaner)
|
||||||
|
tempDir := b.TempDir()
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
outputPath := filepath.Join(tempDir, "large.docx")
|
||||||
|
_ = exporter.Export(course, outputPath)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
b.Run("HTML_Large", func(b *testing.B) {
|
||||||
|
exporter := NewHTMLExporter(htmlCleaner)
|
||||||
|
tempDir := b.TempDir()
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
outputPath := filepath.Join(tempDir, "large.html")
|
||||||
|
_ = exporter.Export(course, outputPath)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// createBenchmarkCourse creates a standard-sized course for benchmarking.
|
||||||
|
func createBenchmarkCourse() *models.Course {
|
||||||
|
return &models.Course{
|
||||||
|
ShareID: "benchmark-id",
|
||||||
|
Author: "Benchmark Author",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
ID: "bench-course",
|
||||||
|
Title: "Benchmark Course",
|
||||||
|
Description: "Performance testing course",
|
||||||
|
NavigationMode: "menu",
|
||||||
|
Lessons: []models.Lesson{
|
||||||
|
{
|
||||||
|
ID: "lesson1",
|
||||||
|
Title: "Introduction",
|
||||||
|
Type: "lesson",
|
||||||
|
Items: []models.Item{
|
||||||
|
{
|
||||||
|
Type: "text",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{
|
||||||
|
Heading: "Welcome",
|
||||||
|
Paragraph: "<p>This is a test paragraph with <strong>HTML</strong> content.</p>",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "list",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{Paragraph: "Item 1"},
|
||||||
|
{Paragraph: "Item 2"},
|
||||||
|
{Paragraph: "Item 3"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// createLargeBenchmarkCourse creates a large course for stress testing.
|
||||||
|
func createLargeBenchmarkCourse() *models.Course {
|
||||||
|
lessons := make([]models.Lesson, 50)
|
||||||
|
for i := range 50 {
|
||||||
|
lessons[i] = models.Lesson{
|
||||||
|
ID: string(rune(i)),
|
||||||
|
Title: "Lesson " + string(rune(i)),
|
||||||
|
Type: "lesson",
|
||||||
|
Description: "<p>This is lesson description with <em>formatting</em>.</p>",
|
||||||
|
Items: []models.Item{
|
||||||
|
{
|
||||||
|
Type: "text",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{
|
||||||
|
Heading: "Section Heading",
|
||||||
|
Paragraph: "<p>Content with <strong>bold</strong> and <em>italic</em> text.</p>",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "list",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{Paragraph: "Point 1"},
|
||||||
|
{Paragraph: "Point 2"},
|
||||||
|
{Paragraph: "Point 3"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "knowledgeCheck",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{
|
||||||
|
Title: "Quiz Question",
|
||||||
|
Answers: []models.Answer{
|
||||||
|
{Title: "Answer A", Correct: false},
|
||||||
|
{Title: "Answer B", Correct: true},
|
||||||
|
{Title: "Answer C", Correct: false},
|
||||||
|
},
|
||||||
|
Feedback: "Good job!",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &models.Course{
|
||||||
|
ShareID: "large-benchmark-id",
|
||||||
|
Author: "Benchmark Author",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
ID: "large-bench-course",
|
||||||
|
Title: "Large Benchmark Course",
|
||||||
|
Description: "Large performance testing course",
|
||||||
|
Lessons: lessons,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -8,11 +8,21 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/fumiama/go-docx"
|
"github.com/fumiama/go-docx"
|
||||||
|
"golang.org/x/text/cases"
|
||||||
|
"golang.org/x/text/language"
|
||||||
|
|
||||||
"github.com/kjanat/articulate-parser/internal/interfaces"
|
"github.com/kjanat/articulate-parser/internal/interfaces"
|
||||||
"github.com/kjanat/articulate-parser/internal/models"
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
"github.com/kjanat/articulate-parser/internal/services"
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Font sizes for DOCX document headings (in half-points, so "32" = 16pt).
|
||||||
|
const (
|
||||||
|
docxTitleSize = "32" // Course title (16pt)
|
||||||
|
docxLessonSize = "28" // Lesson heading (14pt)
|
||||||
|
docxItemSize = "24" // Item heading (12pt)
|
||||||
|
)
|
||||||
|
|
||||||
// DocxExporter implements the Exporter interface for DOCX format.
|
// DocxExporter implements the Exporter interface for DOCX format.
|
||||||
// It converts Articulate Rise course data into a Microsoft Word document
|
// It converts Articulate Rise course data into a Microsoft Word document
|
||||||
// using the go-docx package.
|
// using the go-docx package.
|
||||||
@ -50,7 +60,7 @@ func (e *DocxExporter) Export(course *models.Course, outputPath string) error {
|
|||||||
|
|
||||||
// Add title
|
// Add title
|
||||||
titlePara := doc.AddParagraph()
|
titlePara := doc.AddParagraph()
|
||||||
titlePara.AddText(course.Course.Title).Size("32").Bold()
|
titlePara.AddText(course.Course.Title).Size(docxTitleSize).Bold()
|
||||||
|
|
||||||
// Add description if available
|
// Add description if available
|
||||||
if course.Course.Description != "" {
|
if course.Course.Description != "" {
|
||||||
@ -66,15 +76,24 @@ func (e *DocxExporter) Export(course *models.Course, outputPath string) error {
|
|||||||
|
|
||||||
// Ensure output directory exists and add .docx extension
|
// Ensure output directory exists and add .docx extension
|
||||||
if !strings.HasSuffix(strings.ToLower(outputPath), ".docx") {
|
if !strings.HasSuffix(strings.ToLower(outputPath), ".docx") {
|
||||||
outputPath = outputPath + ".docx"
|
outputPath += ".docx"
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create the file
|
// Create the file
|
||||||
|
// #nosec G304 - Output path is provided by user via CLI argument, which is expected behavior
|
||||||
file, err := os.Create(outputPath)
|
file, err := os.Create(outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to create output file: %w", err)
|
return fmt.Errorf("failed to create output file: %w", err)
|
||||||
}
|
}
|
||||||
defer file.Close()
|
// Ensure file is closed even if WriteTo fails. Close errors are logged but not
|
||||||
|
// fatal since the document content has already been written to disk. A close
|
||||||
|
// error typically indicates a filesystem synchronization issue that doesn't
|
||||||
|
// affect the validity of the exported file.
|
||||||
|
defer func() {
|
||||||
|
if err := file.Close(); err != nil {
|
||||||
|
fmt.Fprintf(os.Stderr, "warning: failed to close output file: %v\n", err)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
// Save the document
|
// Save the document
|
||||||
_, err = doc.WriteTo(file)
|
_, err = doc.WriteTo(file)
|
||||||
@ -94,7 +113,7 @@ func (e *DocxExporter) Export(course *models.Course, outputPath string) error {
|
|||||||
func (e *DocxExporter) exportLesson(doc *docx.Docx, lesson *models.Lesson) {
|
func (e *DocxExporter) exportLesson(doc *docx.Docx, lesson *models.Lesson) {
|
||||||
// Add lesson title
|
// Add lesson title
|
||||||
lessonPara := doc.AddParagraph()
|
lessonPara := doc.AddParagraph()
|
||||||
lessonPara.AddText(fmt.Sprintf("Lesson: %s", lesson.Title)).Size("28").Bold()
|
lessonPara.AddText(fmt.Sprintf("Lesson: %s", lesson.Title)).Size(docxLessonSize).Bold()
|
||||||
|
|
||||||
// Add lesson description if available
|
// Add lesson description if available
|
||||||
if lesson.Description != "" {
|
if lesson.Description != "" {
|
||||||
@ -119,7 +138,8 @@ func (e *DocxExporter) exportItem(doc *docx.Docx, item *models.Item) {
|
|||||||
// Add item type as heading
|
// Add item type as heading
|
||||||
if item.Type != "" {
|
if item.Type != "" {
|
||||||
itemPara := doc.AddParagraph()
|
itemPara := doc.AddParagraph()
|
||||||
itemPara.AddText(strings.Title(item.Type)).Size("24").Bold()
|
caser := cases.Title(language.English)
|
||||||
|
itemPara.AddText(caser.String(item.Type)).Size(docxItemSize).Bold()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add sub-items
|
// Add sub-items
|
||||||
@ -180,10 +200,10 @@ func (e *DocxExporter) exportSubItem(doc *docx.Docx, subItem *models.SubItem) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetSupportedFormat returns the format name this exporter supports.
|
// SupportedFormat returns the format name this exporter supports.
|
||||||
//
|
//
|
||||||
// Returns:
|
// Returns:
|
||||||
// - A string representing the supported format ("docx")
|
// - A string representing the supported format ("docx")
|
||||||
func (e *DocxExporter) GetSupportedFormat() string {
|
func (e *DocxExporter) SupportedFormat() string {
|
||||||
return "docx"
|
return FormatDocx
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,4 +1,3 @@
|
|||||||
// Package exporters_test provides tests for the docx exporter.
|
|
||||||
package exporters
|
package exporters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@ -30,13 +29,13 @@ func TestNewDocxExporter(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestDocxExporter_GetSupportedFormat tests the GetSupportedFormat method.
|
// TestDocxExporter_SupportedFormat tests the SupportedFormat method.
|
||||||
func TestDocxExporter_GetSupportedFormat(t *testing.T) {
|
func TestDocxExporter_SupportedFormat(t *testing.T) {
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
exporter := NewDocxExporter(htmlCleaner)
|
exporter := NewDocxExporter(htmlCleaner)
|
||||||
|
|
||||||
expected := "docx"
|
expected := "docx"
|
||||||
result := exporter.GetSupportedFormat()
|
result := exporter.SupportedFormat()
|
||||||
|
|
||||||
if result != expected {
|
if result != expected {
|
||||||
t.Errorf("Expected format '%s', got '%s'", expected, result)
|
t.Errorf("Expected format '%s', got '%s'", expected, result)
|
||||||
@ -90,7 +89,6 @@ func TestDocxExporter_Export_AddDocxExtension(t *testing.T) {
|
|||||||
|
|
||||||
err := exporter.Export(testCourse, outputPath)
|
err := exporter.Export(testCourse, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -155,7 +153,6 @@ func TestDocxExporter_ExportLesson(t *testing.T) {
|
|||||||
|
|
||||||
err := exporter.Export(course, outputPath)
|
err := exporter.Export(course, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -222,7 +219,6 @@ func TestDocxExporter_ExportItem(t *testing.T) {
|
|||||||
|
|
||||||
err := exporter.Export(course, outputPath)
|
err := exporter.Export(course, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -276,7 +272,6 @@ func TestDocxExporter_ExportSubItem(t *testing.T) {
|
|||||||
|
|
||||||
err := exporter.Export(course, outputPath)
|
err := exporter.Export(course, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -335,7 +330,7 @@ func TestDocxExporter_ComplexCourse(t *testing.T) {
|
|||||||
Caption: "<p>Watch this introductory video</p>",
|
Caption: "<p>Watch this introductory video</p>",
|
||||||
Media: &models.Media{
|
Media: &models.Media{
|
||||||
Video: &models.VideoMedia{
|
Video: &models.VideoMedia{
|
||||||
OriginalUrl: "https://example.com/intro.mp4",
|
OriginalURL: "https://example.com/intro.mp4",
|
||||||
Duration: 300,
|
Duration: 300,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@ -363,7 +358,7 @@ func TestDocxExporter_ComplexCourse(t *testing.T) {
|
|||||||
Caption: "<p>Course overview diagram</p>",
|
Caption: "<p>Course overview diagram</p>",
|
||||||
Media: &models.Media{
|
Media: &models.Media{
|
||||||
Image: &models.ImageMedia{
|
Image: &models.ImageMedia{
|
||||||
OriginalUrl: "https://example.com/overview.png",
|
OriginalURL: "https://example.com/overview.png",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@ -409,7 +404,6 @@ func TestDocxExporter_ComplexCourse(t *testing.T) {
|
|||||||
// Export course
|
// Export course
|
||||||
err := exporter.Export(course, outputPath)
|
err := exporter.Export(course, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -444,7 +438,6 @@ func TestDocxExporter_EmptyCourse(t *testing.T) {
|
|||||||
|
|
||||||
err := exporter.Export(course, outputPath)
|
err := exporter.Export(course, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -493,7 +486,6 @@ func TestDocxExporter_HTMLCleaning(t *testing.T) {
|
|||||||
|
|
||||||
err := exporter.Export(course, outputPath)
|
err := exporter.Export(course, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -516,7 +508,6 @@ func TestDocxExporter_ExistingDocxExtension(t *testing.T) {
|
|||||||
|
|
||||||
err := exporter.Export(testCourse, outputPath)
|
err := exporter.Export(testCourse, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -552,7 +543,6 @@ func TestDocxExporter_CaseInsensitiveExtension(t *testing.T) {
|
|||||||
|
|
||||||
err := exporter.Export(testCourse, outputPath)
|
err := exporter.Export(testCourse, outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
||||||
t.Fatalf("Export failed for case %d (%s): %v", i, testCase, err)
|
t.Fatalf("Export failed for case %d (%s): %v", i, testCase, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -615,12 +605,13 @@ func BenchmarkDocxExporter_Export(b *testing.B) {
|
|||||||
// Create temporary directory
|
// Create temporary directory
|
||||||
tempDir := b.TempDir()
|
tempDir := b.TempDir()
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
outputPath := filepath.Join(tempDir, "benchmark-course.docx")
|
outputPath := filepath.Join(tempDir, "benchmark-course.docx")
|
||||||
_ = exporter.Export(course, outputPath)
|
_ = exporter.Export(course, outputPath)
|
||||||
// Clean up for next iteration
|
// Clean up for next iteration. Remove errors are ignored because we've already
|
||||||
os.Remove(outputPath)
|
// benchmarked the export operation; cleanup failures don't affect the benchmark
|
||||||
|
// measurements or the validity of the next iteration's export.
|
||||||
|
_ = os.Remove(outputPath)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -641,7 +632,7 @@ func BenchmarkDocxExporter_ComplexCourse(b *testing.B) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Fill with test data
|
// Fill with test data
|
||||||
for i := 0; i < 10; i++ {
|
for i := range 10 {
|
||||||
lesson := models.Lesson{
|
lesson := models.Lesson{
|
||||||
ID: "lesson-" + string(rune(i)),
|
ID: "lesson-" + string(rune(i)),
|
||||||
Title: "Lesson " + string(rune(i)),
|
Title: "Lesson " + string(rune(i)),
|
||||||
@ -649,13 +640,13 @@ func BenchmarkDocxExporter_ComplexCourse(b *testing.B) {
|
|||||||
Items: make([]models.Item, 5), // 5 items per lesson
|
Items: make([]models.Item, 5), // 5 items per lesson
|
||||||
}
|
}
|
||||||
|
|
||||||
for j := 0; j < 5; j++ {
|
for j := range 5 {
|
||||||
item := models.Item{
|
item := models.Item{
|
||||||
Type: "text",
|
Type: "text",
|
||||||
Items: make([]models.SubItem, 3), // 3 sub-items per item
|
Items: make([]models.SubItem, 3), // 3 sub-items per item
|
||||||
}
|
}
|
||||||
|
|
||||||
for k := 0; k < 3; k++ {
|
for k := range 3 {
|
||||||
item.Items[k] = models.SubItem{
|
item.Items[k] = models.SubItem{
|
||||||
Heading: "<h3>Heading " + string(rune(k)) + "</h3>",
|
Heading: "<h3>Heading " + string(rune(k)) + "</h3>",
|
||||||
Paragraph: "<p>Paragraph content with <strong>formatting</strong> for performance testing.</p>",
|
Paragraph: "<p>Paragraph content with <strong>formatting</strong> for performance testing.</p>",
|
||||||
@ -670,10 +661,11 @@ func BenchmarkDocxExporter_ComplexCourse(b *testing.B) {
|
|||||||
|
|
||||||
tempDir := b.TempDir()
|
tempDir := b.TempDir()
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
outputPath := filepath.Join(tempDir, "benchmark-complex.docx")
|
outputPath := filepath.Join(tempDir, "benchmark-complex.docx")
|
||||||
_ = exporter.Export(course, outputPath)
|
_ = exporter.Export(course, outputPath)
|
||||||
os.Remove(outputPath)
|
// Remove errors are ignored because we're only benchmarking the export
|
||||||
|
// operation itself; cleanup failures don't affect the benchmark metrics.
|
||||||
|
_ = os.Remove(outputPath)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
101
internal/exporters/example_test.go
Normal file
101
internal/exporters/example_test.go
Normal file
@ -0,0 +1,101 @@
|
|||||||
|
// Package exporters_test provides examples for the exporters package.
|
||||||
|
package exporters_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/exporters"
|
||||||
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ExampleNewFactory demonstrates creating an exporter factory.
|
||||||
|
func ExampleNewFactory() {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
factory := exporters.NewFactory(htmlCleaner)
|
||||||
|
|
||||||
|
// Get supported formats
|
||||||
|
formats := factory.SupportedFormats()
|
||||||
|
fmt.Printf("Supported formats: %d\n", len(formats))
|
||||||
|
// Output: Supported formats: 6
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleFactory_CreateExporter demonstrates creating exporters.
|
||||||
|
func ExampleFactory_CreateExporter() {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
factory := exporters.NewFactory(htmlCleaner)
|
||||||
|
|
||||||
|
// Create a markdown exporter
|
||||||
|
exporter, err := factory.CreateExporter("markdown")
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
fmt.Printf("Created: %s exporter\n", exporter.SupportedFormat())
|
||||||
|
// Output: Created: markdown exporter
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleFactory_CreateExporter_caseInsensitive demonstrates case-insensitive format names.
|
||||||
|
func ExampleFactory_CreateExporter_caseInsensitive() {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
factory := exporters.NewFactory(htmlCleaner)
|
||||||
|
|
||||||
|
// All these work (case-insensitive)
|
||||||
|
formats := []string{"MARKDOWN", "Markdown", "markdown", "MD"}
|
||||||
|
|
||||||
|
for _, format := range formats {
|
||||||
|
exporter, _ := factory.CreateExporter(format)
|
||||||
|
fmt.Printf("%s -> %s\n", format, exporter.SupportedFormat())
|
||||||
|
}
|
||||||
|
// Output:
|
||||||
|
// MARKDOWN -> markdown
|
||||||
|
// Markdown -> markdown
|
||||||
|
// markdown -> markdown
|
||||||
|
// MD -> markdown
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleMarkdownExporter_Export demonstrates exporting to Markdown.
|
||||||
|
func ExampleMarkdownExporter_Export() {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
exporter := exporters.NewMarkdownExporter(htmlCleaner)
|
||||||
|
|
||||||
|
course := &models.Course{
|
||||||
|
ShareID: "example-id",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
Title: "Example Course",
|
||||||
|
Description: "<p>Course description</p>",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
// Export to markdown file
|
||||||
|
err := exporter.Export(course, "output.md")
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
fmt.Println("Export complete")
|
||||||
|
// Output: Export complete
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleDocxExporter_Export demonstrates exporting to DOCX.
|
||||||
|
func ExampleDocxExporter_Export() {
|
||||||
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
|
exporter := exporters.NewDocxExporter(htmlCleaner)
|
||||||
|
|
||||||
|
course := &models.Course{
|
||||||
|
ShareID: "example-id",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
Title: "Example Course",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
// Export to Word document
|
||||||
|
err := exporter.Export(course, "output.docx")
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
fmt.Println("DOCX export complete")
|
||||||
|
// Output: DOCX export complete
|
||||||
|
}
|
||||||
@ -1,5 +1,3 @@
|
|||||||
// Package exporters provides implementations of the Exporter interface
|
|
||||||
// for converting Articulate Rise courses into various file formats.
|
|
||||||
package exporters
|
package exporters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@ -10,6 +8,13 @@ import (
|
|||||||
"github.com/kjanat/articulate-parser/internal/services"
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Format constants for supported export formats.
|
||||||
|
const (
|
||||||
|
FormatMarkdown = "markdown"
|
||||||
|
FormatDocx = "docx"
|
||||||
|
FormatHTML = "html"
|
||||||
|
)
|
||||||
|
|
||||||
// Factory implements the ExporterFactory interface.
|
// Factory implements the ExporterFactory interface.
|
||||||
// It creates appropriate exporter instances based on the requested format.
|
// It creates appropriate exporter instances based on the requested format.
|
||||||
type Factory struct {
|
type Factory struct {
|
||||||
@ -33,33 +38,22 @@ func NewFactory(htmlCleaner *services.HTMLCleaner) interfaces.ExporterFactory {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// CreateExporter creates an exporter for the specified format.
|
// CreateExporter creates an exporter for the specified format.
|
||||||
// It returns an appropriate exporter implementation based on the format string.
|
// Format strings are case-insensitive (e.g., "markdown", "DOCX").
|
||||||
// Format strings are case-insensitive.
|
|
||||||
//
|
|
||||||
// Parameters:
|
|
||||||
// - format: The desired export format (e.g., "markdown", "docx")
|
|
||||||
//
|
|
||||||
// Returns:
|
|
||||||
// - An implementation of the Exporter interface if the format is supported
|
|
||||||
// - An error if the format is not supported
|
|
||||||
func (f *Factory) CreateExporter(format string) (interfaces.Exporter, error) {
|
func (f *Factory) CreateExporter(format string) (interfaces.Exporter, error) {
|
||||||
switch strings.ToLower(format) {
|
switch strings.ToLower(format) {
|
||||||
case "markdown", "md":
|
case FormatMarkdown, "md":
|
||||||
return NewMarkdownExporter(f.htmlCleaner), nil
|
return NewMarkdownExporter(f.htmlCleaner), nil
|
||||||
case "docx", "word":
|
case FormatDocx, "word":
|
||||||
return NewDocxExporter(f.htmlCleaner), nil
|
return NewDocxExporter(f.htmlCleaner), nil
|
||||||
case "html", "htm":
|
case FormatHTML, "htm":
|
||||||
return NewHTMLExporter(f.htmlCleaner), nil
|
return NewHTMLExporter(f.htmlCleaner), nil
|
||||||
default:
|
default:
|
||||||
return nil, fmt.Errorf("unsupported export format: %s", format)
|
return nil, fmt.Errorf("unsupported export format: %s", format)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetSupportedFormats returns a list of all supported export formats.
|
// SupportedFormats returns a list of all supported export formats,
|
||||||
// This includes both primary format names and their aliases.
|
// including both primary format names and their aliases.
|
||||||
//
|
func (f *Factory) SupportedFormats() []string {
|
||||||
// Returns:
|
return []string{FormatMarkdown, "md", FormatDocx, "word", FormatHTML, "htm"}
|
||||||
// - A string slice containing all supported format names
|
|
||||||
func (f *Factory) GetSupportedFormats() []string {
|
|
||||||
return []string{"markdown", "md", "docx", "word", "html", "htm"}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,4 +1,3 @@
|
|||||||
// Package exporters_test provides tests for the exporter factory.
|
|
||||||
package exporters
|
package exporters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@ -125,7 +124,7 @@ func TestFactory_CreateExporter(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Check supported format
|
// Check supported format
|
||||||
supportedFormat := exporter.GetSupportedFormat()
|
supportedFormat := exporter.SupportedFormat()
|
||||||
if supportedFormat != tc.expectedFormat {
|
if supportedFormat != tc.expectedFormat {
|
||||||
t.Errorf("Expected supported format '%s' for format '%s', got '%s'", tc.expectedFormat, tc.format, supportedFormat)
|
t.Errorf("Expected supported format '%s' for format '%s', got '%s'", tc.expectedFormat, tc.format, supportedFormat)
|
||||||
}
|
}
|
||||||
@ -164,7 +163,6 @@ func TestFactory_CreateExporter_CaseInsensitive(t *testing.T) {
|
|||||||
for _, tc := range testCases {
|
for _, tc := range testCases {
|
||||||
t.Run(tc.format, func(t *testing.T) {
|
t.Run(tc.format, func(t *testing.T) {
|
||||||
exporter, err := factory.CreateExporter(tc.format)
|
exporter, err := factory.CreateExporter(tc.format)
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Unexpected error for format '%s': %v", tc.format, err)
|
t.Fatalf("Unexpected error for format '%s': %v", tc.format, err)
|
||||||
}
|
}
|
||||||
@ -173,7 +171,7 @@ func TestFactory_CreateExporter_CaseInsensitive(t *testing.T) {
|
|||||||
t.Fatalf("CreateExporter returned nil for format '%s'", tc.format)
|
t.Fatalf("CreateExporter returned nil for format '%s'", tc.format)
|
||||||
}
|
}
|
||||||
|
|
||||||
supportedFormat := exporter.GetSupportedFormat()
|
supportedFormat := exporter.SupportedFormat()
|
||||||
if supportedFormat != tc.expectedFormat {
|
if supportedFormat != tc.expectedFormat {
|
||||||
t.Errorf("Expected supported format '%s' for format '%s', got '%s'", tc.expectedFormat, tc.format, supportedFormat)
|
t.Errorf("Expected supported format '%s' for format '%s', got '%s'", tc.expectedFormat, tc.format, supportedFormat)
|
||||||
}
|
}
|
||||||
@ -221,15 +219,15 @@ func TestFactory_CreateExporter_ErrorMessages(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestFactory_GetSupportedFormats tests the GetSupportedFormats method.
|
// TestFactory_SupportedFormats tests the SupportedFormats method.
|
||||||
func TestFactory_GetSupportedFormats(t *testing.T) {
|
func TestFactory_SupportedFormats(t *testing.T) {
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
factory := NewFactory(htmlCleaner)
|
factory := NewFactory(htmlCleaner)
|
||||||
|
|
||||||
formats := factory.GetSupportedFormats()
|
formats := factory.SupportedFormats()
|
||||||
|
|
||||||
if formats == nil {
|
if formats == nil {
|
||||||
t.Fatal("GetSupportedFormats() returned nil")
|
t.Fatal("SupportedFormats() returned nil")
|
||||||
}
|
}
|
||||||
|
|
||||||
expected := []string{"markdown", "md", "docx", "word", "html", "htm"}
|
expected := []string{"markdown", "md", "docx", "word", "html", "htm"}
|
||||||
@ -246,22 +244,22 @@ func TestFactory_GetSupportedFormats(t *testing.T) {
|
|||||||
for _, format := range formats {
|
for _, format := range formats {
|
||||||
exporter, err := factory.CreateExporter(format)
|
exporter, err := factory.CreateExporter(format)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Errorf("Format '%s' from GetSupportedFormats() should be creatable, got error: %v", format, err)
|
t.Errorf("Format '%s' from SupportedFormats() should be creatable, got error: %v", format, err)
|
||||||
}
|
}
|
||||||
if exporter == nil {
|
if exporter == nil {
|
||||||
t.Errorf("Format '%s' from GetSupportedFormats() should create non-nil exporter", format)
|
t.Errorf("Format '%s' from SupportedFormats() should create non-nil exporter", format)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestFactory_GetSupportedFormats_Immutable tests that the returned slice is safe to modify.
|
// TestFactory_SupportedFormats_Immutable tests that the returned slice is safe to modify.
|
||||||
func TestFactory_GetSupportedFormats_Immutable(t *testing.T) {
|
func TestFactory_SupportedFormats_Immutable(t *testing.T) {
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
factory := NewFactory(htmlCleaner)
|
factory := NewFactory(htmlCleaner)
|
||||||
|
|
||||||
// Get formats twice
|
// Get formats twice
|
||||||
formats1 := factory.GetSupportedFormats()
|
formats1 := factory.SupportedFormats()
|
||||||
formats2 := factory.GetSupportedFormats()
|
formats2 := factory.SupportedFormats()
|
||||||
|
|
||||||
// Modify first slice
|
// Modify first slice
|
||||||
if len(formats1) > 0 {
|
if len(formats1) > 0 {
|
||||||
@ -270,13 +268,13 @@ func TestFactory_GetSupportedFormats_Immutable(t *testing.T) {
|
|||||||
|
|
||||||
// Check that second call returns unmodified data
|
// Check that second call returns unmodified data
|
||||||
if len(formats2) > 0 && formats2[0] == "modified" {
|
if len(formats2) > 0 && formats2[0] == "modified" {
|
||||||
t.Error("GetSupportedFormats() should return independent slices")
|
t.Error("SupportedFormats() should return independent slices")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Verify original functionality still works
|
// Verify original functionality still works
|
||||||
formats3 := factory.GetSupportedFormats()
|
formats3 := factory.SupportedFormats()
|
||||||
if len(formats3) == 0 {
|
if len(formats3) == 0 {
|
||||||
t.Error("GetSupportedFormats() should still return formats after modification")
|
t.Error("SupportedFormats() should still return formats after modification")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -436,7 +434,7 @@ func TestFactory_FormatNormalization(t *testing.T) {
|
|||||||
t.Fatalf("Failed to create exporter for '%s': %v", tc.input, err)
|
t.Fatalf("Failed to create exporter for '%s': %v", tc.input, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
format := exporter.GetSupportedFormat()
|
format := exporter.SupportedFormat()
|
||||||
if format != tc.expected {
|
if format != tc.expected {
|
||||||
t.Errorf("Expected format '%s' for input '%s', got '%s'", tc.expected, tc.input, format)
|
t.Errorf("Expected format '%s' for input '%s', got '%s'", tc.expected, tc.input, format)
|
||||||
}
|
}
|
||||||
@ -449,8 +447,7 @@ func BenchmarkFactory_CreateExporter(b *testing.B) {
|
|||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
factory := NewFactory(htmlCleaner)
|
factory := NewFactory(htmlCleaner)
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
_, _ = factory.CreateExporter("markdown")
|
_, _ = factory.CreateExporter("markdown")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -460,19 +457,17 @@ func BenchmarkFactory_CreateExporter_Docx(b *testing.B) {
|
|||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
factory := NewFactory(htmlCleaner)
|
factory := NewFactory(htmlCleaner)
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
_, _ = factory.CreateExporter("docx")
|
_, _ = factory.CreateExporter("docx")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// BenchmarkFactory_GetSupportedFormats benchmarks the GetSupportedFormats method.
|
// BenchmarkFactory_SupportedFormats benchmarks the SupportedFormats method.
|
||||||
func BenchmarkFactory_GetSupportedFormats(b *testing.B) {
|
func BenchmarkFactory_SupportedFormats(b *testing.B) {
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
factory := NewFactory(htmlCleaner)
|
factory := NewFactory(htmlCleaner)
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
_ = factory.SupportedFormats()
|
||||||
_ = factory.GetSupportedFormats()
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,24 +1,30 @@
|
|||||||
// Package exporters provides implementations of the Exporter interface
|
|
||||||
// for converting Articulate Rise courses into various file formats.
|
|
||||||
package exporters
|
package exporters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
_ "embed"
|
||||||
"fmt"
|
"fmt"
|
||||||
"html"
|
"html/template"
|
||||||
|
"io"
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
|
||||||
|
|
||||||
"github.com/kjanat/articulate-parser/internal/interfaces"
|
"github.com/kjanat/articulate-parser/internal/interfaces"
|
||||||
"github.com/kjanat/articulate-parser/internal/models"
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
"github.com/kjanat/articulate-parser/internal/services"
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
//go:embed html_styles.css
|
||||||
|
var defaultCSS string
|
||||||
|
|
||||||
|
//go:embed html_template.gohtml
|
||||||
|
var htmlTemplate string
|
||||||
|
|
||||||
// HTMLExporter implements the Exporter interface for HTML format.
|
// HTMLExporter implements the Exporter interface for HTML format.
|
||||||
// It converts Articulate Rise course data into a structured HTML document.
|
// It converts Articulate Rise course data into a structured HTML document using templates.
|
||||||
type HTMLExporter struct {
|
type HTMLExporter struct {
|
||||||
// htmlCleaner is used to convert HTML content to plain text when needed
|
// htmlCleaner is used to convert HTML content to plain text when needed
|
||||||
htmlCleaner *services.HTMLCleaner
|
htmlCleaner *services.HTMLCleaner
|
||||||
|
// tmpl holds the parsed HTML template
|
||||||
|
tmpl *template.Template
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewHTMLExporter creates a new HTMLExporter instance.
|
// NewHTMLExporter creates a new HTMLExporter instance.
|
||||||
@ -30,8 +36,21 @@ type HTMLExporter struct {
|
|||||||
// Returns:
|
// Returns:
|
||||||
// - An implementation of the Exporter interface for HTML format
|
// - An implementation of the Exporter interface for HTML format
|
||||||
func NewHTMLExporter(htmlCleaner *services.HTMLCleaner) interfaces.Exporter {
|
func NewHTMLExporter(htmlCleaner *services.HTMLCleaner) interfaces.Exporter {
|
||||||
|
// Parse the template with custom functions
|
||||||
|
funcMap := template.FuncMap{
|
||||||
|
"safeHTML": func(s string) template.HTML {
|
||||||
|
return template.HTML(s) // #nosec G203 - HTML content is from trusted course data
|
||||||
|
},
|
||||||
|
"safeCSS": func(s string) template.CSS {
|
||||||
|
return template.CSS(s) // #nosec G203 - CSS content is from trusted embedded file
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
tmpl := template.Must(template.New("html").Funcs(funcMap).Parse(htmlTemplate))
|
||||||
|
|
||||||
return &HTMLExporter{
|
return &HTMLExporter{
|
||||||
htmlCleaner: htmlCleaner,
|
htmlCleaner: htmlCleaner,
|
||||||
|
tmpl: tmpl,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -46,431 +65,50 @@ func NewHTMLExporter(htmlCleaner *services.HTMLCleaner) interfaces.Exporter {
|
|||||||
// Returns:
|
// Returns:
|
||||||
// - An error if writing to the output file fails
|
// - An error if writing to the output file fails
|
||||||
func (e *HTMLExporter) Export(course *models.Course, outputPath string) error {
|
func (e *HTMLExporter) Export(course *models.Course, outputPath string) error {
|
||||||
var buf bytes.Buffer
|
f, err := os.Create(outputPath)
|
||||||
|
if err != nil {
|
||||||
// Write HTML document structure
|
return fmt.Errorf("failed to create file: %w", err)
|
||||||
buf.WriteString("<!DOCTYPE html>\n")
|
|
||||||
buf.WriteString("<html lang=\"en\">\n")
|
|
||||||
buf.WriteString("<head>\n")
|
|
||||||
buf.WriteString(" <meta charset=\"UTF-8\">\n")
|
|
||||||
buf.WriteString(" <meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0\">\n")
|
|
||||||
buf.WriteString(fmt.Sprintf(" <title>%s</title>\n", html.EscapeString(course.Course.Title)))
|
|
||||||
buf.WriteString(" <style>\n")
|
|
||||||
buf.WriteString(e.getDefaultCSS())
|
|
||||||
buf.WriteString(" </style>\n")
|
|
||||||
buf.WriteString("</head>\n")
|
|
||||||
buf.WriteString("<body>\n")
|
|
||||||
|
|
||||||
// Write course header
|
|
||||||
buf.WriteString(fmt.Sprintf(" <header>\n <h1>%s</h1>\n", html.EscapeString(course.Course.Title)))
|
|
||||||
|
|
||||||
if course.Course.Description != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <div class=\"course-description\">%s</div>\n", course.Course.Description))
|
|
||||||
}
|
}
|
||||||
buf.WriteString(" </header>\n\n")
|
defer func() {
|
||||||
|
// Close errors are logged but not fatal since the content has already been written.
|
||||||
// Add metadata section
|
// The file must be closed to flush buffers, but a close error doesn't invalidate
|
||||||
buf.WriteString(" <section class=\"course-info\">\n")
|
// the data already written to disk.
|
||||||
buf.WriteString(" <h2>Course Information</h2>\n")
|
if closeErr := f.Close(); closeErr != nil {
|
||||||
buf.WriteString(" <ul>\n")
|
// Note: In production, this should log via a logger passed to the exporter.
|
||||||
buf.WriteString(fmt.Sprintf(" <li><strong>Course ID:</strong> %s</li>\n", html.EscapeString(course.Course.ID)))
|
// For now, we silently ignore close errors as they're non-fatal.
|
||||||
buf.WriteString(fmt.Sprintf(" <li><strong>Share ID:</strong> %s</li>\n", html.EscapeString(course.ShareID)))
|
_ = closeErr
|
||||||
buf.WriteString(fmt.Sprintf(" <li><strong>Navigation Mode:</strong> %s</li>\n", html.EscapeString(course.Course.NavigationMode)))
|
|
||||||
if course.Course.ExportSettings != nil {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <li><strong>Export Format:</strong> %s</li>\n", html.EscapeString(course.Course.ExportSettings.Format)))
|
|
||||||
}
|
|
||||||
buf.WriteString(" </ul>\n")
|
|
||||||
buf.WriteString(" </section>\n\n")
|
|
||||||
|
|
||||||
// Process lessons
|
|
||||||
lessonCounter := 0
|
|
||||||
for _, lesson := range course.Course.Lessons {
|
|
||||||
if lesson.Type == "section" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <section class=\"course-section\">\n <h2>%s</h2>\n </section>\n\n", html.EscapeString(lesson.Title)))
|
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
lessonCounter++
|
return e.WriteHTML(f, course)
|
||||||
buf.WriteString(fmt.Sprintf(" <section class=\"lesson\">\n <h3>Lesson %d: %s</h3>\n", lessonCounter, html.EscapeString(lesson.Title)))
|
|
||||||
|
|
||||||
if lesson.Description != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <div class=\"lesson-description\">%s</div>\n", lesson.Description))
|
|
||||||
}
|
|
||||||
|
|
||||||
// Process lesson items
|
|
||||||
for _, item := range lesson.Items {
|
|
||||||
e.processItemToHTML(&buf, item)
|
|
||||||
}
|
|
||||||
|
|
||||||
buf.WriteString(" </section>\n\n")
|
|
||||||
}
|
|
||||||
|
|
||||||
buf.WriteString("</body>\n")
|
|
||||||
buf.WriteString("</html>\n")
|
|
||||||
|
|
||||||
return os.WriteFile(outputPath, buf.Bytes(), 0644)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetSupportedFormat returns the format name this exporter supports
|
// WriteHTML writes the HTML content to an io.Writer.
|
||||||
|
// This allows for better testability and flexibility in output destinations.
|
||||||
|
//
|
||||||
|
// Parameters:
|
||||||
|
// - w: The writer to output HTML content to
|
||||||
|
// - course: The course data model to export
|
||||||
|
//
|
||||||
|
// Returns:
|
||||||
|
// - An error if writing fails
|
||||||
|
func (e *HTMLExporter) WriteHTML(w io.Writer, course *models.Course) error {
|
||||||
|
// Prepare template data
|
||||||
|
data := prepareTemplateData(course, e.htmlCleaner)
|
||||||
|
|
||||||
|
// Execute template
|
||||||
|
if err := e.tmpl.Execute(w, data); err != nil {
|
||||||
|
return fmt.Errorf("failed to execute template: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// SupportedFormat returns the format name this exporter supports
|
||||||
// It indicates the file format that the HTMLExporter can generate.
|
// It indicates the file format that the HTMLExporter can generate.
|
||||||
//
|
//
|
||||||
// Returns:
|
// Returns:
|
||||||
// - A string representing the supported format ("html")
|
// - A string representing the supported format ("html")
|
||||||
func (e *HTMLExporter) GetSupportedFormat() string {
|
func (e *HTMLExporter) SupportedFormat() string {
|
||||||
return "html"
|
return FormatHTML
|
||||||
}
|
|
||||||
|
|
||||||
// getDefaultCSS returns basic CSS styling for the HTML document
|
|
||||||
func (e *HTMLExporter) getDefaultCSS() string {
|
|
||||||
return `
|
|
||||||
body {
|
|
||||||
font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, sans-serif;
|
|
||||||
line-height: 1.6;
|
|
||||||
color: #333;
|
|
||||||
max-width: 800px;
|
|
||||||
margin: 0 auto;
|
|
||||||
padding: 20px;
|
|
||||||
background-color: #f9f9f9;
|
|
||||||
}
|
|
||||||
header {
|
|
||||||
background: linear-gradient(135deg, #667eea 0%, #764ba2 100%);
|
|
||||||
color: white;
|
|
||||||
padding: 2rem;
|
|
||||||
border-radius: 10px;
|
|
||||||
margin-bottom: 2rem;
|
|
||||||
box-shadow: 0 4px 6px rgba(0, 0, 0, 0.1);
|
|
||||||
}
|
|
||||||
header h1 {
|
|
||||||
margin: 0;
|
|
||||||
font-size: 2.5rem;
|
|
||||||
font-weight: 300;
|
|
||||||
}
|
|
||||||
.course-description {
|
|
||||||
margin-top: 1rem;
|
|
||||||
font-size: 1.1rem;
|
|
||||||
opacity: 0.9;
|
|
||||||
}
|
|
||||||
.course-info {
|
|
||||||
background: white;
|
|
||||||
padding: 1.5rem;
|
|
||||||
border-radius: 8px;
|
|
||||||
margin-bottom: 2rem;
|
|
||||||
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.1);
|
|
||||||
}
|
|
||||||
.course-info h2 {
|
|
||||||
margin-top: 0;
|
|
||||||
color: #4a5568;
|
|
||||||
border-bottom: 2px solid #e2e8f0;
|
|
||||||
padding-bottom: 0.5rem;
|
|
||||||
}
|
|
||||||
.course-info ul {
|
|
||||||
list-style: none;
|
|
||||||
padding: 0;
|
|
||||||
}
|
|
||||||
.course-info li {
|
|
||||||
margin: 0.5rem 0;
|
|
||||||
padding: 0.5rem;
|
|
||||||
background: #f7fafc;
|
|
||||||
border-radius: 4px;
|
|
||||||
}
|
|
||||||
.course-section {
|
|
||||||
background: #4299e1;
|
|
||||||
color: white;
|
|
||||||
padding: 1.5rem;
|
|
||||||
border-radius: 8px;
|
|
||||||
margin: 2rem 0;
|
|
||||||
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.1);
|
|
||||||
}
|
|
||||||
.course-section h2 {
|
|
||||||
margin: 0;
|
|
||||||
font-weight: 400;
|
|
||||||
}
|
|
||||||
.lesson {
|
|
||||||
background: white;
|
|
||||||
padding: 2rem;
|
|
||||||
border-radius: 8px;
|
|
||||||
margin: 2rem 0;
|
|
||||||
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.1);
|
|
||||||
border-left: 4px solid #4299e1;
|
|
||||||
}
|
|
||||||
.lesson h3 {
|
|
||||||
margin-top: 0;
|
|
||||||
color: #2d3748;
|
|
||||||
font-size: 1.5rem;
|
|
||||||
}
|
|
||||||
.lesson-description {
|
|
||||||
margin: 1rem 0;
|
|
||||||
padding: 1rem;
|
|
||||||
background: #f7fafc;
|
|
||||||
border-radius: 4px;
|
|
||||||
border-left: 3px solid #4299e1;
|
|
||||||
}
|
|
||||||
.item {
|
|
||||||
margin: 1.5rem 0;
|
|
||||||
padding: 1rem;
|
|
||||||
border-radius: 6px;
|
|
||||||
background: #fafafa;
|
|
||||||
border: 1px solid #e2e8f0;
|
|
||||||
}
|
|
||||||
.item h4 {
|
|
||||||
margin-top: 0;
|
|
||||||
color: #4a5568;
|
|
||||||
font-size: 1.2rem;
|
|
||||||
text-transform: capitalize;
|
|
||||||
}
|
|
||||||
.text-item {
|
|
||||||
background: #f0fff4;
|
|
||||||
border-left: 3px solid #48bb78;
|
|
||||||
}
|
|
||||||
.list-item {
|
|
||||||
background: #fffaf0;
|
|
||||||
border-left: 3px solid #ed8936;
|
|
||||||
}
|
|
||||||
.knowledge-check {
|
|
||||||
background: #e6fffa;
|
|
||||||
border-left: 3px solid #38b2ac;
|
|
||||||
}
|
|
||||||
.multimedia-item {
|
|
||||||
background: #faf5ff;
|
|
||||||
border-left: 3px solid #9f7aea;
|
|
||||||
}
|
|
||||||
.interactive-item {
|
|
||||||
background: #fff5f5;
|
|
||||||
border-left: 3px solid #f56565;
|
|
||||||
}
|
|
||||||
.unknown-item {
|
|
||||||
background: #f7fafc;
|
|
||||||
border-left: 3px solid #a0aec0;
|
|
||||||
}
|
|
||||||
.answers {
|
|
||||||
margin: 1rem 0;
|
|
||||||
}
|
|
||||||
.answers h5 {
|
|
||||||
margin: 0.5rem 0;
|
|
||||||
color: #4a5568;
|
|
||||||
}
|
|
||||||
.answers ol {
|
|
||||||
margin: 0.5rem 0;
|
|
||||||
padding-left: 1.5rem;
|
|
||||||
}
|
|
||||||
.answers li {
|
|
||||||
margin: 0.3rem 0;
|
|
||||||
padding: 0.3rem;
|
|
||||||
}
|
|
||||||
.correct-answer {
|
|
||||||
background: #c6f6d5;
|
|
||||||
border-radius: 3px;
|
|
||||||
font-weight: bold;
|
|
||||||
}
|
|
||||||
.correct-answer::after {
|
|
||||||
content: " ✓";
|
|
||||||
color: #38a169;
|
|
||||||
}
|
|
||||||
.feedback {
|
|
||||||
margin: 1rem 0;
|
|
||||||
padding: 1rem;
|
|
||||||
background: #edf2f7;
|
|
||||||
border-radius: 4px;
|
|
||||||
border-left: 3px solid #4299e1;
|
|
||||||
font-style: italic;
|
|
||||||
}
|
|
||||||
.media-info {
|
|
||||||
background: #edf2f7;
|
|
||||||
padding: 1rem;
|
|
||||||
border-radius: 4px;
|
|
||||||
margin: 0.5rem 0;
|
|
||||||
}
|
|
||||||
.media-info strong {
|
|
||||||
color: #4a5568;
|
|
||||||
}
|
|
||||||
hr {
|
|
||||||
border: none;
|
|
||||||
height: 2px;
|
|
||||||
background: linear-gradient(to right, #667eea, #764ba2);
|
|
||||||
margin: 2rem 0;
|
|
||||||
border-radius: 1px;
|
|
||||||
}
|
|
||||||
ul {
|
|
||||||
padding-left: 1.5rem;
|
|
||||||
}
|
|
||||||
li {
|
|
||||||
margin: 0.5rem 0;
|
|
||||||
}
|
|
||||||
`
|
|
||||||
}
|
|
||||||
|
|
||||||
// processItemToHTML converts a course item into HTML format
|
|
||||||
// and appends it to the provided buffer. It handles different item types
|
|
||||||
// with appropriate HTML formatting.
|
|
||||||
//
|
|
||||||
// Parameters:
|
|
||||||
// - buf: The buffer to write the HTML content to
|
|
||||||
// - item: The course item to process
|
|
||||||
func (e *HTMLExporter) processItemToHTML(buf *bytes.Buffer, item models.Item) {
|
|
||||||
switch strings.ToLower(item.Type) {
|
|
||||||
case "text":
|
|
||||||
e.processTextItem(buf, item)
|
|
||||||
case "list":
|
|
||||||
e.processListItem(buf, item)
|
|
||||||
case "knowledgecheck":
|
|
||||||
e.processKnowledgeCheckItem(buf, item)
|
|
||||||
case "multimedia":
|
|
||||||
e.processMultimediaItem(buf, item)
|
|
||||||
case "image":
|
|
||||||
e.processImageItem(buf, item)
|
|
||||||
case "interactive":
|
|
||||||
e.processInteractiveItem(buf, item)
|
|
||||||
case "divider":
|
|
||||||
e.processDividerItem(buf)
|
|
||||||
default:
|
|
||||||
e.processUnknownItem(buf, item)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// processTextItem handles text content with headings and paragraphs
|
|
||||||
func (e *HTMLExporter) processTextItem(buf *bytes.Buffer, item models.Item) {
|
|
||||||
buf.WriteString(" <div class=\"item text-item\">\n")
|
|
||||||
buf.WriteString(" <h4>Text Content</h4>\n")
|
|
||||||
for _, subItem := range item.Items {
|
|
||||||
if subItem.Heading != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <h5>%s</h5>\n", subItem.Heading))
|
|
||||||
}
|
|
||||||
if subItem.Paragraph != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <div>%s</div>\n", subItem.Paragraph))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
buf.WriteString(" </div>\n\n")
|
|
||||||
}
|
|
||||||
|
|
||||||
// processListItem handles list content
|
|
||||||
func (e *HTMLExporter) processListItem(buf *bytes.Buffer, item models.Item) {
|
|
||||||
buf.WriteString(" <div class=\"item list-item\">\n")
|
|
||||||
buf.WriteString(" <h4>List</h4>\n")
|
|
||||||
buf.WriteString(" <ul>\n")
|
|
||||||
for _, subItem := range item.Items {
|
|
||||||
if subItem.Paragraph != "" {
|
|
||||||
cleanText := e.htmlCleaner.CleanHTML(subItem.Paragraph)
|
|
||||||
buf.WriteString(fmt.Sprintf(" <li>%s</li>\n", html.EscapeString(cleanText)))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
buf.WriteString(" </ul>\n")
|
|
||||||
buf.WriteString(" </div>\n\n")
|
|
||||||
}
|
|
||||||
|
|
||||||
// processKnowledgeCheckItem handles quiz questions and answers
|
|
||||||
func (e *HTMLExporter) processKnowledgeCheckItem(buf *bytes.Buffer, item models.Item) {
|
|
||||||
buf.WriteString(" <div class=\"item knowledge-check\">\n")
|
|
||||||
buf.WriteString(" <h4>Knowledge Check</h4>\n")
|
|
||||||
for _, subItem := range item.Items {
|
|
||||||
if subItem.Title != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <p><strong>Question:</strong> %s</p>\n", subItem.Title))
|
|
||||||
}
|
|
||||||
if len(subItem.Answers) > 0 {
|
|
||||||
e.processAnswers(buf, subItem.Answers)
|
|
||||||
}
|
|
||||||
if subItem.Feedback != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <div class=\"feedback\"><strong>Feedback:</strong> %s</div>\n", subItem.Feedback))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
buf.WriteString(" </div>\n\n")
|
|
||||||
}
|
|
||||||
|
|
||||||
// processMultimediaItem handles multimedia content like videos
|
|
||||||
func (e *HTMLExporter) processMultimediaItem(buf *bytes.Buffer, item models.Item) {
|
|
||||||
buf.WriteString(" <div class=\"item multimedia-item\">\n")
|
|
||||||
buf.WriteString(" <h4>Media Content</h4>\n")
|
|
||||||
for _, subItem := range item.Items {
|
|
||||||
if subItem.Title != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <h5>%s</h5>\n", subItem.Title))
|
|
||||||
}
|
|
||||||
if subItem.Media != nil {
|
|
||||||
if subItem.Media.Video != nil {
|
|
||||||
buf.WriteString(" <div class=\"media-info\">\n")
|
|
||||||
buf.WriteString(fmt.Sprintf(" <p><strong>Video:</strong> %s</p>\n", html.EscapeString(subItem.Media.Video.OriginalUrl)))
|
|
||||||
if subItem.Media.Video.Duration > 0 {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <p><strong>Duration:</strong> %d seconds</p>\n", subItem.Media.Video.Duration))
|
|
||||||
}
|
|
||||||
buf.WriteString(" </div>\n")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if subItem.Caption != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <div><em>%s</em></div>\n", subItem.Caption))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
buf.WriteString(" </div>\n\n")
|
|
||||||
}
|
|
||||||
|
|
||||||
// processImageItem handles image content
|
|
||||||
func (e *HTMLExporter) processImageItem(buf *bytes.Buffer, item models.Item) {
|
|
||||||
buf.WriteString(" <div class=\"item multimedia-item\">\n")
|
|
||||||
buf.WriteString(" <h4>Image</h4>\n")
|
|
||||||
for _, subItem := range item.Items {
|
|
||||||
if subItem.Media != nil && subItem.Media.Image != nil {
|
|
||||||
buf.WriteString(" <div class=\"media-info\">\n")
|
|
||||||
buf.WriteString(fmt.Sprintf(" <p><strong>Image:</strong> %s</p>\n", html.EscapeString(subItem.Media.Image.OriginalUrl)))
|
|
||||||
buf.WriteString(" </div>\n")
|
|
||||||
}
|
|
||||||
if subItem.Caption != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <div><em>%s</em></div>\n", subItem.Caption))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
buf.WriteString(" </div>\n\n")
|
|
||||||
}
|
|
||||||
|
|
||||||
// processInteractiveItem handles interactive content
|
|
||||||
func (e *HTMLExporter) processInteractiveItem(buf *bytes.Buffer, item models.Item) {
|
|
||||||
buf.WriteString(" <div class=\"item interactive-item\">\n")
|
|
||||||
buf.WriteString(" <h4>Interactive Content</h4>\n")
|
|
||||||
for _, subItem := range item.Items {
|
|
||||||
if subItem.Title != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <p><strong>%s</strong></p>\n", subItem.Title))
|
|
||||||
}
|
|
||||||
if subItem.Paragraph != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <div>%s</div>\n", subItem.Paragraph))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
buf.WriteString(" </div>\n\n")
|
|
||||||
}
|
|
||||||
|
|
||||||
// processDividerItem handles divider elements
|
|
||||||
func (e *HTMLExporter) processDividerItem(buf *bytes.Buffer) {
|
|
||||||
buf.WriteString(" <hr>\n\n")
|
|
||||||
}
|
|
||||||
|
|
||||||
// processUnknownItem handles unknown or unsupported item types
|
|
||||||
func (e *HTMLExporter) processUnknownItem(buf *bytes.Buffer, item models.Item) {
|
|
||||||
if len(item.Items) > 0 {
|
|
||||||
buf.WriteString(" <div class=\"item unknown-item\">\n")
|
|
||||||
buf.WriteString(fmt.Sprintf(" <h4>%s Content</h4>\n", strings.Title(item.Type)))
|
|
||||||
for _, subItem := range item.Items {
|
|
||||||
e.processGenericSubItem(buf, subItem)
|
|
||||||
}
|
|
||||||
buf.WriteString(" </div>\n\n")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// processGenericSubItem processes sub-items for unknown types
|
|
||||||
func (e *HTMLExporter) processGenericSubItem(buf *bytes.Buffer, subItem models.SubItem) {
|
|
||||||
if subItem.Title != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <p><strong>%s</strong></p>\n", subItem.Title))
|
|
||||||
}
|
|
||||||
if subItem.Paragraph != "" {
|
|
||||||
buf.WriteString(fmt.Sprintf(" <div>%s</div>\n", subItem.Paragraph))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// processAnswers processes answer choices for quiz questions
|
|
||||||
func (e *HTMLExporter) processAnswers(buf *bytes.Buffer, answers []models.Answer) {
|
|
||||||
buf.WriteString(" <div class=\"answers\">\n")
|
|
||||||
buf.WriteString(" <h5>Answers:</h5>\n")
|
|
||||||
buf.WriteString(" <ol>\n")
|
|
||||||
for _, answer := range answers {
|
|
||||||
cssClass := ""
|
|
||||||
if answer.Correct {
|
|
||||||
cssClass = " class=\"correct-answer\""
|
|
||||||
}
|
|
||||||
buf.WriteString(fmt.Sprintf(" <li%s>%s</li>\n", cssClass, html.EscapeString(answer.Title)))
|
|
||||||
}
|
|
||||||
buf.WriteString(" </ol>\n")
|
|
||||||
buf.WriteString(" </div>\n")
|
|
||||||
}
|
}
|
||||||
|
|||||||
175
internal/exporters/html_styles.css
Normal file
175
internal/exporters/html_styles.css
Normal file
@ -0,0 +1,175 @@
|
|||||||
|
body {
|
||||||
|
font-family:
|
||||||
|
-apple-system, BlinkMacSystemFont, "Segoe UI", Roboto, Oxygen, Ubuntu,
|
||||||
|
Cantarell, sans-serif;
|
||||||
|
line-height: 1.6;
|
||||||
|
color: #333;
|
||||||
|
max-width: 800px;
|
||||||
|
margin: 0 auto;
|
||||||
|
padding: 20px;
|
||||||
|
background-color: #f9f9f9;
|
||||||
|
}
|
||||||
|
header {
|
||||||
|
background: linear-gradient(135deg, #667eea 0%, #764ba2 100%);
|
||||||
|
color: white;
|
||||||
|
padding: 2rem;
|
||||||
|
border-radius: 10px;
|
||||||
|
margin-bottom: 2rem;
|
||||||
|
box-shadow: 0 4px 6px rgba(0, 0, 0, 0.1);
|
||||||
|
}
|
||||||
|
header h1 {
|
||||||
|
margin: 0;
|
||||||
|
font-size: 2.5rem;
|
||||||
|
font-weight: 300;
|
||||||
|
}
|
||||||
|
.course-description {
|
||||||
|
margin-top: 1rem;
|
||||||
|
font-size: 1.1rem;
|
||||||
|
opacity: 0.9;
|
||||||
|
}
|
||||||
|
.course-info {
|
||||||
|
background: white;
|
||||||
|
padding: 1.5rem;
|
||||||
|
border-radius: 8px;
|
||||||
|
margin-bottom: 2rem;
|
||||||
|
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.1);
|
||||||
|
}
|
||||||
|
.course-info h2 {
|
||||||
|
margin-top: 0;
|
||||||
|
color: #4a5568;
|
||||||
|
border-bottom: 2px solid #e2e8f0;
|
||||||
|
padding-bottom: 0.5rem;
|
||||||
|
}
|
||||||
|
.course-info ul {
|
||||||
|
list-style: none;
|
||||||
|
padding: 0;
|
||||||
|
}
|
||||||
|
.course-info li {
|
||||||
|
margin: 0.5rem 0;
|
||||||
|
padding: 0.5rem;
|
||||||
|
background: #f7fafc;
|
||||||
|
border-radius: 4px;
|
||||||
|
}
|
||||||
|
.course-section {
|
||||||
|
background: #4299e1;
|
||||||
|
color: white;
|
||||||
|
padding: 1.5rem;
|
||||||
|
border-radius: 8px;
|
||||||
|
margin: 2rem 0;
|
||||||
|
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.1);
|
||||||
|
}
|
||||||
|
.course-section h2 {
|
||||||
|
margin: 0;
|
||||||
|
font-weight: 400;
|
||||||
|
}
|
||||||
|
.lesson {
|
||||||
|
background: white;
|
||||||
|
padding: 2rem;
|
||||||
|
border-radius: 8px;
|
||||||
|
margin: 2rem 0;
|
||||||
|
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.1);
|
||||||
|
border-left: 4px solid #4299e1;
|
||||||
|
}
|
||||||
|
.lesson h3 {
|
||||||
|
margin-top: 0;
|
||||||
|
color: #2d3748;
|
||||||
|
font-size: 1.5rem;
|
||||||
|
}
|
||||||
|
.lesson-description {
|
||||||
|
margin: 1rem 0;
|
||||||
|
padding: 1rem;
|
||||||
|
background: #f7fafc;
|
||||||
|
border-radius: 4px;
|
||||||
|
border-left: 3px solid #4299e1;
|
||||||
|
}
|
||||||
|
.item {
|
||||||
|
margin: 1.5rem 0;
|
||||||
|
padding: 1rem;
|
||||||
|
border-radius: 6px;
|
||||||
|
background: #fafafa;
|
||||||
|
border: 1px solid #e2e8f0;
|
||||||
|
}
|
||||||
|
.item h4 {
|
||||||
|
margin-top: 0;
|
||||||
|
color: #4a5568;
|
||||||
|
font-size: 1.2rem;
|
||||||
|
text-transform: capitalize;
|
||||||
|
}
|
||||||
|
.text-item {
|
||||||
|
background: #f0fff4;
|
||||||
|
border-left: 3px solid #48bb78;
|
||||||
|
}
|
||||||
|
.list-item {
|
||||||
|
background: #fffaf0;
|
||||||
|
border-left: 3px solid #ed8936;
|
||||||
|
}
|
||||||
|
.knowledge-check {
|
||||||
|
background: #e6fffa;
|
||||||
|
border-left: 3px solid #38b2ac;
|
||||||
|
}
|
||||||
|
.multimedia-item {
|
||||||
|
background: #faf5ff;
|
||||||
|
border-left: 3px solid #9f7aea;
|
||||||
|
}
|
||||||
|
.interactive-item {
|
||||||
|
background: #fff5f5;
|
||||||
|
border-left: 3px solid #f56565;
|
||||||
|
}
|
||||||
|
.unknown-item {
|
||||||
|
background: #f7fafc;
|
||||||
|
border-left: 3px solid #a0aec0;
|
||||||
|
}
|
||||||
|
.answers {
|
||||||
|
margin: 1rem 0;
|
||||||
|
}
|
||||||
|
.answers h5 {
|
||||||
|
margin: 0.5rem 0;
|
||||||
|
color: #4a5568;
|
||||||
|
}
|
||||||
|
.answers ol {
|
||||||
|
margin: 0.5rem 0;
|
||||||
|
padding-left: 1.5rem;
|
||||||
|
}
|
||||||
|
.answers li {
|
||||||
|
margin: 0.3rem 0;
|
||||||
|
padding: 0.3rem;
|
||||||
|
}
|
||||||
|
.correct-answer {
|
||||||
|
background: #c6f6d5;
|
||||||
|
border-radius: 3px;
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
.correct-answer::after {
|
||||||
|
content: " ✓";
|
||||||
|
color: #38a169;
|
||||||
|
}
|
||||||
|
.feedback {
|
||||||
|
margin: 1rem 0;
|
||||||
|
padding: 1rem;
|
||||||
|
background: #edf2f7;
|
||||||
|
border-radius: 4px;
|
||||||
|
border-left: 3px solid #4299e1;
|
||||||
|
font-style: italic;
|
||||||
|
}
|
||||||
|
.media-info {
|
||||||
|
background: #edf2f7;
|
||||||
|
padding: 1rem;
|
||||||
|
border-radius: 4px;
|
||||||
|
margin: 0.5rem 0;
|
||||||
|
}
|
||||||
|
.media-info strong {
|
||||||
|
color: #4a5568;
|
||||||
|
}
|
||||||
|
hr {
|
||||||
|
border: none;
|
||||||
|
height: 2px;
|
||||||
|
background: linear-gradient(to right, #667eea, #764ba2);
|
||||||
|
margin: 2rem 0;
|
||||||
|
border-radius: 1px;
|
||||||
|
}
|
||||||
|
ul {
|
||||||
|
padding-left: 1.5rem;
|
||||||
|
}
|
||||||
|
li {
|
||||||
|
margin: 0.5rem 0;
|
||||||
|
}
|
||||||
183
internal/exporters/html_template.gohtml
Normal file
183
internal/exporters/html_template.gohtml
Normal file
@ -0,0 +1,183 @@
|
|||||||
|
<!DOCTYPE html>
|
||||||
|
<html lang="en">
|
||||||
|
<head>
|
||||||
|
<meta charset="UTF-8">
|
||||||
|
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||||
|
<title>{{.Course.Title}}</title>
|
||||||
|
<style>
|
||||||
|
{{safeCSS .CSS}}
|
||||||
|
</style>
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<header>
|
||||||
|
<h1>{{.Course.Title}}</h1>
|
||||||
|
{{if .Course.Description}}
|
||||||
|
<div class="course-description">{{safeHTML .Course.Description}}</div>
|
||||||
|
{{end}}
|
||||||
|
</header>
|
||||||
|
|
||||||
|
<section class="course-info">
|
||||||
|
<h2>Course Information</h2>
|
||||||
|
<ul>
|
||||||
|
<li><strong>Course ID:</strong> {{.Course.ID}}</li>
|
||||||
|
<li><strong>Share ID:</strong> {{.ShareID}}</li>
|
||||||
|
<li><strong>Navigation Mode:</strong> {{.Course.NavigationMode}}</li>
|
||||||
|
{{if .Course.ExportSettings}}
|
||||||
|
<li><strong>Export Format:</strong> {{.Course.ExportSettings.Format}}</li>
|
||||||
|
{{end}}
|
||||||
|
</ul>
|
||||||
|
</section>
|
||||||
|
|
||||||
|
{{range .Sections}}
|
||||||
|
{{if eq .Type "section"}}
|
||||||
|
<section class="course-section">
|
||||||
|
<h2>{{.Title}}</h2>
|
||||||
|
</section>
|
||||||
|
{{else}}
|
||||||
|
<section class="lesson">
|
||||||
|
<h3>Lesson {{.Number}}: {{.Title}}</h3>
|
||||||
|
{{if .Description}}
|
||||||
|
<div class="lesson-description">{{safeHTML .Description}}</div>
|
||||||
|
{{end}}
|
||||||
|
{{range .Items}}
|
||||||
|
{{template "item" .}}
|
||||||
|
{{end}}
|
||||||
|
</section>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
{{define "item"}}
|
||||||
|
{{if eq .Type "text"}}{{template "textItem" .}}
|
||||||
|
{{else if eq .Type "list"}}{{template "listItem" .}}
|
||||||
|
{{else if eq .Type "knowledgecheck"}}{{template "knowledgeCheckItem" .}}
|
||||||
|
{{else if eq .Type "multimedia"}}{{template "multimediaItem" .}}
|
||||||
|
{{else if eq .Type "image"}}{{template "imageItem" .}}
|
||||||
|
{{else if eq .Type "interactive"}}{{template "interactiveItem" .}}
|
||||||
|
{{else if eq .Type "divider"}}{{template "dividerItem" .}}
|
||||||
|
{{else}}{{template "unknownItem" .}}
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
{{define "textItem"}}
|
||||||
|
<div class="item text-item">
|
||||||
|
<h4>Text Content</h4>
|
||||||
|
{{range .Items}}
|
||||||
|
{{if .Heading}}
|
||||||
|
{{safeHTML .Heading}}
|
||||||
|
{{end}}
|
||||||
|
{{if .Paragraph}}
|
||||||
|
<div>{{safeHTML .Paragraph}}</div>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
{{define "listItem"}}
|
||||||
|
<div class="item list-item">
|
||||||
|
<h4>List</h4>
|
||||||
|
<ul>
|
||||||
|
{{range .Items}}
|
||||||
|
{{if .Paragraph}}
|
||||||
|
<li>{{.CleanText}}</li>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
</ul>
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
{{define "knowledgeCheckItem"}}
|
||||||
|
<div class="item knowledge-check">
|
||||||
|
<h4>Knowledge Check</h4>
|
||||||
|
{{range .Items}}
|
||||||
|
{{if .Title}}
|
||||||
|
<p><strong>Question:</strong> {{safeHTML .Title}}</p>
|
||||||
|
{{end}}
|
||||||
|
{{if .Answers}}
|
||||||
|
<div class="answers">
|
||||||
|
<h5>Answers:</h5>
|
||||||
|
<ol>
|
||||||
|
{{range .Answers}}
|
||||||
|
<li{{if .Correct}} class="correct-answer"{{end}}>{{.Title}}</li>
|
||||||
|
{{end}}
|
||||||
|
</ol>
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
{{if .Feedback}}
|
||||||
|
<div class="feedback"><strong>Feedback:</strong> {{safeHTML .Feedback}}</div>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
{{define "multimediaItem"}}
|
||||||
|
<div class="item multimedia-item">
|
||||||
|
<h4>Media Content</h4>
|
||||||
|
{{range .Items}}
|
||||||
|
{{if .Title}}
|
||||||
|
<h5>{{.Title}}</h5>
|
||||||
|
{{end}}
|
||||||
|
{{if .Media}}
|
||||||
|
{{if .Media.Video}}
|
||||||
|
<div class="media-info">
|
||||||
|
<p><strong>Video:</strong> {{.Media.Video.OriginalURL}}</p>
|
||||||
|
{{if gt .Media.Video.Duration 0}}
|
||||||
|
<p><strong>Duration:</strong> {{.Media.Video.Duration}} seconds</p>
|
||||||
|
{{end}}
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
{{if .Caption}}
|
||||||
|
<div><em>{{.Caption}}</em></div>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
{{define "imageItem"}}
|
||||||
|
<div class="item multimedia-item">
|
||||||
|
<h4>Image</h4>
|
||||||
|
{{range .Items}}
|
||||||
|
{{if and .Media .Media.Image}}
|
||||||
|
<div class="media-info">
|
||||||
|
<p><strong>Image:</strong> {{.Media.Image.OriginalURL}}</p>
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
{{if .Caption}}
|
||||||
|
<div><em>{{.Caption}}</em></div>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
{{define "interactiveItem"}}
|
||||||
|
<div class="item interactive-item">
|
||||||
|
<h4>Interactive Content</h4>
|
||||||
|
{{range .Items}}
|
||||||
|
{{if .Title}}
|
||||||
|
<p><strong>{{.Title}}</strong></p>
|
||||||
|
{{end}}
|
||||||
|
{{if .Paragraph}}
|
||||||
|
<div>{{safeHTML .Paragraph}}</div>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
{{define "dividerItem"}}
|
||||||
|
<hr>
|
||||||
|
{{end}}
|
||||||
|
|
||||||
|
{{define "unknownItem"}}
|
||||||
|
<div class="item unknown-item">
|
||||||
|
<h4>{{.TypeTitle}} Content</h4>
|
||||||
|
{{range .Items}}
|
||||||
|
{{if .Title}}
|
||||||
|
<p><strong>{{.Title}}</strong></p>
|
||||||
|
{{end}}
|
||||||
|
{{if .Paragraph}}
|
||||||
|
<div>{{safeHTML .Paragraph}}</div>
|
||||||
|
{{end}}
|
||||||
|
{{end}}
|
||||||
|
</div>
|
||||||
|
{{end}}
|
||||||
131
internal/exporters/html_template_data.go
Normal file
131
internal/exporters/html_template_data.go
Normal file
@ -0,0 +1,131 @@
|
|||||||
|
package exporters
|
||||||
|
|
||||||
|
import (
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"golang.org/x/text/cases"
|
||||||
|
"golang.org/x/text/language"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Item type constants.
|
||||||
|
const (
|
||||||
|
itemTypeText = "text"
|
||||||
|
itemTypeList = "list"
|
||||||
|
itemTypeKnowledgeCheck = "knowledgecheck"
|
||||||
|
itemTypeMultimedia = "multimedia"
|
||||||
|
itemTypeImage = "image"
|
||||||
|
itemTypeInteractive = "interactive"
|
||||||
|
itemTypeDivider = "divider"
|
||||||
|
)
|
||||||
|
|
||||||
|
// templateData represents the data structure passed to the HTML template.
|
||||||
|
type templateData struct {
|
||||||
|
Course models.CourseInfo
|
||||||
|
ShareID string
|
||||||
|
Sections []templateSection
|
||||||
|
CSS string
|
||||||
|
}
|
||||||
|
|
||||||
|
// templateSection represents a course section or lesson.
|
||||||
|
type templateSection struct {
|
||||||
|
Type string
|
||||||
|
Title string
|
||||||
|
Number int
|
||||||
|
Description string
|
||||||
|
Items []templateItem
|
||||||
|
}
|
||||||
|
|
||||||
|
// templateItem represents a course item with preprocessed data.
|
||||||
|
type templateItem struct {
|
||||||
|
Type string
|
||||||
|
TypeTitle string
|
||||||
|
Items []templateSubItem
|
||||||
|
}
|
||||||
|
|
||||||
|
// templateSubItem represents a sub-item with preprocessed data.
|
||||||
|
type templateSubItem struct {
|
||||||
|
Heading string
|
||||||
|
Paragraph string
|
||||||
|
Title string
|
||||||
|
Caption string
|
||||||
|
CleanText string
|
||||||
|
Answers []models.Answer
|
||||||
|
Feedback string
|
||||||
|
Media *models.Media
|
||||||
|
}
|
||||||
|
|
||||||
|
// prepareTemplateData converts a Course model into template-friendly data.
|
||||||
|
func prepareTemplateData(course *models.Course, htmlCleaner *services.HTMLCleaner) *templateData {
|
||||||
|
data := &templateData{
|
||||||
|
Course: course.Course,
|
||||||
|
ShareID: course.ShareID,
|
||||||
|
Sections: make([]templateSection, 0, len(course.Course.Lessons)),
|
||||||
|
CSS: defaultCSS,
|
||||||
|
}
|
||||||
|
|
||||||
|
lessonCounter := 0
|
||||||
|
for _, lesson := range course.Course.Lessons {
|
||||||
|
section := templateSection{
|
||||||
|
Type: lesson.Type,
|
||||||
|
Title: lesson.Title,
|
||||||
|
Description: lesson.Description,
|
||||||
|
}
|
||||||
|
|
||||||
|
if lesson.Type != "section" {
|
||||||
|
lessonCounter++
|
||||||
|
section.Number = lessonCounter
|
||||||
|
section.Items = prepareItems(lesson.Items, htmlCleaner)
|
||||||
|
}
|
||||||
|
|
||||||
|
data.Sections = append(data.Sections, section)
|
||||||
|
}
|
||||||
|
|
||||||
|
return data
|
||||||
|
}
|
||||||
|
|
||||||
|
// prepareItems converts model Items to template Items.
|
||||||
|
func prepareItems(items []models.Item, htmlCleaner *services.HTMLCleaner) []templateItem {
|
||||||
|
result := make([]templateItem, 0, len(items))
|
||||||
|
|
||||||
|
for _, item := range items {
|
||||||
|
tItem := templateItem{
|
||||||
|
Type: strings.ToLower(item.Type),
|
||||||
|
Items: make([]templateSubItem, 0, len(item.Items)),
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set type title for unknown items
|
||||||
|
if tItem.Type != itemTypeText && tItem.Type != itemTypeList && tItem.Type != itemTypeKnowledgeCheck &&
|
||||||
|
tItem.Type != itemTypeMultimedia && tItem.Type != itemTypeImage && tItem.Type != itemTypeInteractive &&
|
||||||
|
tItem.Type != itemTypeDivider {
|
||||||
|
caser := cases.Title(language.English)
|
||||||
|
tItem.TypeTitle = caser.String(item.Type)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process sub-items
|
||||||
|
for _, subItem := range item.Items {
|
||||||
|
tSubItem := templateSubItem{
|
||||||
|
Heading: subItem.Heading,
|
||||||
|
Paragraph: subItem.Paragraph,
|
||||||
|
Title: subItem.Title,
|
||||||
|
Caption: subItem.Caption,
|
||||||
|
Answers: subItem.Answers,
|
||||||
|
Feedback: subItem.Feedback,
|
||||||
|
Media: subItem.Media,
|
||||||
|
}
|
||||||
|
|
||||||
|
// Clean HTML for list items
|
||||||
|
if tItem.Type == itemTypeList && subItem.Paragraph != "" {
|
||||||
|
tSubItem.CleanText = htmlCleaner.CleanHTML(subItem.Paragraph)
|
||||||
|
}
|
||||||
|
|
||||||
|
tItem.Items = append(tItem.Items, tSubItem)
|
||||||
|
}
|
||||||
|
|
||||||
|
result = append(result, tItem)
|
||||||
|
}
|
||||||
|
|
||||||
|
return result
|
||||||
|
}
|
||||||
@ -1,8 +1,6 @@
|
|||||||
// Package exporters_test provides tests for the html exporter.
|
|
||||||
package exporters
|
package exporters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
@ -30,15 +28,19 @@ func TestNewHTMLExporter(t *testing.T) {
|
|||||||
if htmlExporter.htmlCleaner == nil {
|
if htmlExporter.htmlCleaner == nil {
|
||||||
t.Error("htmlCleaner should not be nil")
|
t.Error("htmlCleaner should not be nil")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if htmlExporter.tmpl == nil {
|
||||||
|
t.Error("template should not be nil")
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestHTMLExporter_GetSupportedFormat tests the GetSupportedFormat method.
|
// TestHTMLExporter_SupportedFormat tests the SupportedFormat method.
|
||||||
func TestHTMLExporter_GetSupportedFormat(t *testing.T) {
|
func TestHTMLExporter_SupportedFormat(t *testing.T) {
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
exporter := NewHTMLExporter(htmlCleaner)
|
exporter := NewHTMLExporter(htmlCleaner)
|
||||||
|
|
||||||
expected := "html"
|
expected := "html"
|
||||||
result := exporter.GetSupportedFormat()
|
result := exporter.SupportedFormat()
|
||||||
|
|
||||||
if result != expected {
|
if result != expected {
|
||||||
t.Errorf("Expected format '%s', got '%s'", expected, result)
|
t.Errorf("Expected format '%s', got '%s'", expected, result)
|
||||||
@ -119,6 +121,7 @@ func TestHTMLExporter_Export(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if !strings.Contains(contentStr, "font-family") {
|
if !strings.Contains(contentStr, "font-family") {
|
||||||
|
t.Logf("Generated HTML (first 500 chars):\n%s", contentStr[:min(500, len(contentStr))])
|
||||||
t.Error("Output should contain CSS font-family")
|
t.Error("Output should contain CSS font-family")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -139,409 +142,7 @@ func TestHTMLExporter_Export_InvalidPath(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessTextItem tests the processTextItem method.
|
// TestHTMLExporter_ComplexCourse tests export of a course with complex content.
|
||||||
func TestHTMLExporter_ProcessTextItem(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
item := models.Item{
|
|
||||||
Type: "text",
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{
|
|
||||||
Heading: "<h1>Test Heading</h1>",
|
|
||||||
Paragraph: "<p>Test paragraph with <strong>bold</strong> text.</p>",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Paragraph: "<p>Another paragraph.</p>",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processTextItem(&buf, item)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
|
|
||||||
if !strings.Contains(result, "text-item") {
|
|
||||||
t.Error("Should contain text-item CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Text Content") {
|
|
||||||
t.Error("Should contain text content heading")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<h1>Test Heading</h1>") {
|
|
||||||
t.Error("Should preserve HTML heading")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<strong>bold</strong>") {
|
|
||||||
t.Error("Should preserve HTML formatting in paragraph")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessListItem tests the processListItem method.
|
|
||||||
func TestHTMLExporter_ProcessListItem(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
item := models.Item{
|
|
||||||
Type: "list",
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{Paragraph: "<p>First item</p>"},
|
|
||||||
{Paragraph: "<p>Second item with <em>emphasis</em></p>"},
|
|
||||||
{Paragraph: "<p>Third item</p>"},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processListItem(&buf, item)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
|
|
||||||
if !strings.Contains(result, "list-item") {
|
|
||||||
t.Error("Should contain list-item CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<ul>") {
|
|
||||||
t.Error("Should contain unordered list")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<li>First item</li>") {
|
|
||||||
t.Error("Should contain first list item")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<li>Second item with emphasis</li>") {
|
|
||||||
t.Error("Should contain second list item with cleaned HTML")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<li>Third item</li>") {
|
|
||||||
t.Error("Should contain third list item")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessKnowledgeCheckItem tests the processKnowledgeCheckItem method.
|
|
||||||
func TestHTMLExporter_ProcessKnowledgeCheckItem(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
item := models.Item{
|
|
||||||
Type: "knowledgeCheck",
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{
|
|
||||||
Title: "<p>What is the correct answer?</p>",
|
|
||||||
Answers: []models.Answer{
|
|
||||||
{Title: "Wrong answer", Correct: false},
|
|
||||||
{Title: "Correct answer", Correct: true},
|
|
||||||
{Title: "Another wrong answer", Correct: false},
|
|
||||||
},
|
|
||||||
Feedback: "<p>Great job! This is the feedback.</p>",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processKnowledgeCheckItem(&buf, item)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
|
|
||||||
if !strings.Contains(result, "knowledge-check") {
|
|
||||||
t.Error("Should contain knowledge-check CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Knowledge Check") {
|
|
||||||
t.Error("Should contain knowledge check heading")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "What is the correct answer?") {
|
|
||||||
t.Error("Should contain question text")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Wrong answer") {
|
|
||||||
t.Error("Should contain first answer")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "correct-answer") {
|
|
||||||
t.Error("Should mark correct answer with CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Feedback") {
|
|
||||||
t.Error("Should contain feedback section")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessMultimediaItem tests the processMultimediaItem method.
|
|
||||||
func TestHTMLExporter_ProcessMultimediaItem(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
item := models.Item{
|
|
||||||
Type: "multimedia",
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{
|
|
||||||
Title: "<p>Video Title</p>",
|
|
||||||
Media: &models.Media{
|
|
||||||
Video: &models.VideoMedia{
|
|
||||||
OriginalUrl: "https://example.com/video.mp4",
|
|
||||||
Duration: 120,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
Caption: "<p>Video caption</p>",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processMultimediaItem(&buf, item)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
|
|
||||||
if !strings.Contains(result, "multimedia-item") {
|
|
||||||
t.Error("Should contain multimedia-item CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Media Content") {
|
|
||||||
t.Error("Should contain media content heading")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Video Title") {
|
|
||||||
t.Error("Should contain video title")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "https://example.com/video.mp4") {
|
|
||||||
t.Error("Should contain video URL")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "120 seconds") {
|
|
||||||
t.Error("Should contain video duration")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Video caption") {
|
|
||||||
t.Error("Should contain video caption")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessImageItem tests the processImageItem method.
|
|
||||||
func TestHTMLExporter_ProcessImageItem(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
item := models.Item{
|
|
||||||
Type: "image",
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{
|
|
||||||
Media: &models.Media{
|
|
||||||
Image: &models.ImageMedia{
|
|
||||||
OriginalUrl: "https://example.com/image.png",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
Caption: "<p>Image caption</p>",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processImageItem(&buf, item)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
|
|
||||||
if !strings.Contains(result, "multimedia-item") {
|
|
||||||
t.Error("Should contain multimedia-item CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Image") {
|
|
||||||
t.Error("Should contain image heading")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "https://example.com/image.png") {
|
|
||||||
t.Error("Should contain image URL")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Image caption") {
|
|
||||||
t.Error("Should contain image caption")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessInteractiveItem tests the processInteractiveItem method.
|
|
||||||
func TestHTMLExporter_ProcessInteractiveItem(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
item := models.Item{
|
|
||||||
Type: "interactive",
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{
|
|
||||||
Title: "<p>Interactive element title</p>",
|
|
||||||
Paragraph: "<p>Interactive content description</p>",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processInteractiveItem(&buf, item)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
|
|
||||||
if !strings.Contains(result, "interactive-item") {
|
|
||||||
t.Error("Should contain interactive-item CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Interactive Content") {
|
|
||||||
t.Error("Should contain interactive content heading")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Interactive element title") {
|
|
||||||
t.Error("Should contain interactive element title")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Interactive content description") {
|
|
||||||
t.Error("Should contain interactive content description")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessDividerItem tests the processDividerItem method.
|
|
||||||
func TestHTMLExporter_ProcessDividerItem(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
exporter.processDividerItem(&buf)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
expected := " <hr>\n\n"
|
|
||||||
|
|
||||||
if result != expected {
|
|
||||||
t.Errorf("Expected %q, got %q", expected, result)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessUnknownItem tests the processUnknownItem method.
|
|
||||||
func TestHTMLExporter_ProcessUnknownItem(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
item := models.Item{
|
|
||||||
Type: "unknown",
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{
|
|
||||||
Title: "<p>Unknown item title</p>",
|
|
||||||
Paragraph: "<p>Unknown item content</p>",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processUnknownItem(&buf, item)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
|
|
||||||
if !strings.Contains(result, "unknown-item") {
|
|
||||||
t.Error("Should contain unknown-item CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Unknown Content") {
|
|
||||||
t.Error("Should contain unknown content heading")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Unknown item title") {
|
|
||||||
t.Error("Should contain unknown item title")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "Unknown item content") {
|
|
||||||
t.Error("Should contain unknown item content")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessAnswers tests the processAnswers method.
|
|
||||||
func TestHTMLExporter_ProcessAnswers(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
var buf bytes.Buffer
|
|
||||||
answers := []models.Answer{
|
|
||||||
{Title: "Answer 1", Correct: false},
|
|
||||||
{Title: "Answer 2", Correct: true},
|
|
||||||
{Title: "Answer 3", Correct: false},
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processAnswers(&buf, answers)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
|
|
||||||
if !strings.Contains(result, "answers") {
|
|
||||||
t.Error("Should contain answers CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<h5>Answers:</h5>") {
|
|
||||||
t.Error("Should contain answers heading")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<ol>") {
|
|
||||||
t.Error("Should contain ordered list")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<li>Answer 1</li>") {
|
|
||||||
t.Error("Should contain first answer")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "correct-answer") {
|
|
||||||
t.Error("Should mark correct answer with CSS class")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<li class=\"correct-answer\">Answer 2</li>") {
|
|
||||||
t.Error("Should mark correct answer properly")
|
|
||||||
}
|
|
||||||
if !strings.Contains(result, "<li>Answer 3</li>") {
|
|
||||||
t.Error("Should contain third answer")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ProcessItemToHTML_AllTypes tests all item types.
|
|
||||||
func TestHTMLExporter_ProcessItemToHTML_AllTypes(t *testing.T) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
itemType string
|
|
||||||
expectedText string
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "text item",
|
|
||||||
itemType: "text",
|
|
||||||
expectedText: "Text Content",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "list item",
|
|
||||||
itemType: "list",
|
|
||||||
expectedText: "List",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "knowledge check item",
|
|
||||||
itemType: "knowledgeCheck",
|
|
||||||
expectedText: "Knowledge Check",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "multimedia item",
|
|
||||||
itemType: "multimedia",
|
|
||||||
expectedText: "Media Content",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "image item",
|
|
||||||
itemType: "image",
|
|
||||||
expectedText: "Image",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "interactive item",
|
|
||||||
itemType: "interactive",
|
|
||||||
expectedText: "Interactive Content",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "divider item",
|
|
||||||
itemType: "divider",
|
|
||||||
expectedText: "<hr>",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "unknown item",
|
|
||||||
itemType: "unknown",
|
|
||||||
expectedText: "Unknown Content",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
var buf bytes.Buffer
|
|
||||||
item := models.Item{
|
|
||||||
Type: tt.itemType,
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{Title: "Test title", Paragraph: "Test content"},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
// Handle empty unknown items
|
|
||||||
if tt.itemType == "unknown" && tt.expectedText == "" {
|
|
||||||
item.Items = []models.SubItem{}
|
|
||||||
}
|
|
||||||
|
|
||||||
exporter.processItemToHTML(&buf, item)
|
|
||||||
|
|
||||||
result := buf.String()
|
|
||||||
if tt.expectedText != "" && !strings.Contains(result, tt.expectedText) {
|
|
||||||
t.Errorf("Expected content to contain: %q\nGot: %q", tt.expectedText, result)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHTMLExporter_ComplexCourse tests export of a complex course structure.
|
|
||||||
func TestHTMLExporter_ComplexCourse(t *testing.T) {
|
func TestHTMLExporter_ComplexCourse(t *testing.T) {
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
exporter := NewHTMLExporter(htmlCleaner)
|
exporter := NewHTMLExporter(htmlCleaner)
|
||||||
@ -743,11 +344,17 @@ func TestHTMLExporter_HTMLCleaning(t *testing.T) {
|
|||||||
Type: "text",
|
Type: "text",
|
||||||
Items: []models.SubItem{
|
Items: []models.SubItem{
|
||||||
{
|
{
|
||||||
Heading: "<h1>Heading with <em>emphasis</em> and & entities</h1>",
|
Heading: "<h2>HTML Heading</h2>",
|
||||||
Paragraph: "<p>Paragraph with <code> entities and <strong>formatting</strong>.</p>",
|
Paragraph: "<p>Content with <em>emphasis</em> and <strong>strong</strong> text.</p>",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
Type: "list",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{Paragraph: "<p>List item with <b>bold</b> text</p>"},
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@ -762,13 +369,6 @@ func TestHTMLExporter_HTMLCleaning(t *testing.T) {
|
|||||||
t.Fatalf("Export failed: %v", err)
|
t.Fatalf("Export failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Verify file was created (basic check that HTML handling didn't break export)
|
|
||||||
if _, err := os.Stat(outputPath); os.IsNotExist(err) {
|
|
||||||
t.Fatal("Output file was not created")
|
|
||||||
}
|
|
||||||
|
|
||||||
// Read content and verify some HTML is preserved (descriptions, headings, paragraphs)
|
|
||||||
// while list items are cleaned for safety
|
|
||||||
content, err := os.ReadFile(outputPath)
|
content, err := os.ReadFile(outputPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Failed to read output file: %v", err)
|
t.Fatalf("Failed to read output file: %v", err)
|
||||||
@ -776,19 +376,23 @@ func TestHTMLExporter_HTMLCleaning(t *testing.T) {
|
|||||||
|
|
||||||
contentStr := string(content)
|
contentStr := string(content)
|
||||||
|
|
||||||
// HTML should be preserved in some places
|
// HTML content in descriptions should be preserved
|
||||||
if !strings.Contains(contentStr, "<b>bold</b>") {
|
if !strings.Contains(contentStr, "<b>bold</b>") {
|
||||||
t.Error("Should preserve HTML formatting in descriptions")
|
t.Error("Should preserve HTML formatting in descriptions")
|
||||||
}
|
}
|
||||||
if !strings.Contains(contentStr, "<h1>Heading with <em>emphasis</em>") {
|
|
||||||
|
// HTML content in headings should be preserved
|
||||||
|
if !strings.Contains(contentStr, "<h2>HTML Heading</h2>") {
|
||||||
t.Error("Should preserve HTML in headings")
|
t.Error("Should preserve HTML in headings")
|
||||||
}
|
}
|
||||||
if !strings.Contains(contentStr, "<strong>formatting</strong>") {
|
|
||||||
t.Error("Should preserve HTML in paragraphs")
|
// List items should have HTML tags stripped (cleaned)
|
||||||
|
if !strings.Contains(contentStr, "List item with bold text") {
|
||||||
|
t.Error("Should clean HTML from list items")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// createTestCourseForHTML creates a test course for HTML export testing.
|
// createTestCourseForHTML creates a test course for HTML export tests.
|
||||||
func createTestCourseForHTML() *models.Course {
|
func createTestCourseForHTML() *models.Course {
|
||||||
return &models.Course{
|
return &models.Course{
|
||||||
ShareID: "test-share-id",
|
ShareID: "test-share-id",
|
||||||
@ -838,37 +442,13 @@ func BenchmarkHTMLExporter_Export(b *testing.B) {
|
|||||||
exporter := NewHTMLExporter(htmlCleaner)
|
exporter := NewHTMLExporter(htmlCleaner)
|
||||||
course := createTestCourseForHTML()
|
course := createTestCourseForHTML()
|
||||||
|
|
||||||
// Create temporary directory
|
|
||||||
tempDir := b.TempDir()
|
tempDir := b.TempDir()
|
||||||
|
|
||||||
b.ResetTimer()
|
for i := range b.N {
|
||||||
for i := 0; i < b.N; i++ {
|
outputPath := filepath.Join(tempDir, "bench-course-"+string(rune(i))+".html")
|
||||||
outputPath := filepath.Join(tempDir, "benchmark-course.html")
|
if err := exporter.Export(course, outputPath); err != nil {
|
||||||
_ = exporter.Export(course, outputPath)
|
b.Fatalf("Export failed: %v", err)
|
||||||
// Clean up for next iteration
|
}
|
||||||
os.Remove(outputPath)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// BenchmarkHTMLExporter_ProcessTextItem benchmarks text item processing.
|
|
||||||
func BenchmarkHTMLExporter_ProcessTextItem(b *testing.B) {
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
|
||||||
exporter := &HTMLExporter{htmlCleaner: htmlCleaner}
|
|
||||||
|
|
||||||
item := models.Item{
|
|
||||||
Type: "text",
|
|
||||||
Items: []models.SubItem{
|
|
||||||
{
|
|
||||||
Heading: "<h1>Benchmark Heading</h1>",
|
|
||||||
Paragraph: "<p>Benchmark paragraph with <strong>formatting</strong>.</p>",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
b.ResetTimer()
|
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
var buf bytes.Buffer
|
|
||||||
exporter.processTextItem(&buf, item)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -889,39 +469,40 @@ func BenchmarkHTMLExporter_ComplexCourse(b *testing.B) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Fill with test data
|
// Fill with test data
|
||||||
for i := 0; i < 10; i++ {
|
for i := range 10 {
|
||||||
lesson := models.Lesson{
|
lesson := models.Lesson{
|
||||||
ID: "lesson-" + string(rune(i)),
|
ID: "lesson-" + string(rune(i)),
|
||||||
Title: "Lesson " + string(rune(i)),
|
Title: "Benchmark Lesson " + string(rune(i)),
|
||||||
Type: "lesson",
|
Type: "lesson",
|
||||||
Items: make([]models.Item, 5), // 5 items per lesson
|
Description: "<p>Lesson description</p>",
|
||||||
|
Items: []models.Item{
|
||||||
|
{
|
||||||
|
Type: "text",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{
|
||||||
|
Heading: "<h2>Heading</h2>",
|
||||||
|
Paragraph: "<p>Paragraph with content.</p>",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "list",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{Paragraph: "<p>Item 1</p>"},
|
||||||
|
{Paragraph: "<p>Item 2</p>"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for j := 0; j < 5; j++ {
|
|
||||||
item := models.Item{
|
|
||||||
Type: "text",
|
|
||||||
Items: make([]models.SubItem, 3), // 3 sub-items per item
|
|
||||||
}
|
|
||||||
|
|
||||||
for k := 0; k < 3; k++ {
|
|
||||||
item.Items[k] = models.SubItem{
|
|
||||||
Heading: "<h3>Heading " + string(rune(k)) + "</h3>",
|
|
||||||
Paragraph: "<p>Paragraph content with <strong>formatting</strong> for performance testing.</p>",
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
lesson.Items[j] = item
|
|
||||||
}
|
|
||||||
|
|
||||||
course.Course.Lessons[i] = lesson
|
course.Course.Lessons[i] = lesson
|
||||||
}
|
}
|
||||||
|
|
||||||
tempDir := b.TempDir()
|
tempDir := b.TempDir()
|
||||||
|
|
||||||
b.ResetTimer()
|
for i := range b.N {
|
||||||
for i := 0; i < b.N; i++ {
|
outputPath := filepath.Join(tempDir, "bench-complex-"+string(rune(i))+".html")
|
||||||
outputPath := filepath.Join(tempDir, "benchmark-complex.html")
|
if err := exporter.Export(course, outputPath); err != nil {
|
||||||
_ = exporter.Export(course, outputPath)
|
b.Fatalf("Export failed: %v", err)
|
||||||
os.Remove(outputPath)
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,5 +1,3 @@
|
|||||||
// Package exporters provides implementations of the Exporter interface
|
|
||||||
// for converting Articulate Rise courses into various file formats.
|
|
||||||
package exporters
|
package exporters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@ -8,6 +6,9 @@ import (
|
|||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"golang.org/x/text/cases"
|
||||||
|
"golang.org/x/text/language"
|
||||||
|
|
||||||
"github.com/kjanat/articulate-parser/internal/interfaces"
|
"github.com/kjanat/articulate-parser/internal/interfaces"
|
||||||
"github.com/kjanat/articulate-parser/internal/models"
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
"github.com/kjanat/articulate-parser/internal/services"
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
@ -34,16 +35,7 @@ func NewMarkdownExporter(htmlCleaner *services.HTMLCleaner) interfaces.Exporter
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Export exports a course to Markdown format.
|
// Export converts the course to Markdown format and writes it to the output path.
|
||||||
// It generates a structured Markdown document from the course data
|
|
||||||
// and writes it to the specified output path.
|
|
||||||
//
|
|
||||||
// Parameters:
|
|
||||||
// - course: The course data model to export
|
|
||||||
// - outputPath: The file path where the Markdown content will be written
|
|
||||||
//
|
|
||||||
// Returns:
|
|
||||||
// - An error if writing to the output file fails
|
|
||||||
func (e *MarkdownExporter) Export(course *models.Course, outputPath string) error {
|
func (e *MarkdownExporter) Export(course *models.Course, outputPath string) error {
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
|
|
||||||
@ -87,30 +79,27 @@ func (e *MarkdownExporter) Export(course *models.Course, outputPath string) erro
|
|||||||
buf.WriteString("\n---\n\n")
|
buf.WriteString("\n---\n\n")
|
||||||
}
|
}
|
||||||
|
|
||||||
return os.WriteFile(outputPath, buf.Bytes(), 0644)
|
// #nosec G306 - 0644 is appropriate for export files that should be readable by others
|
||||||
|
if err := os.WriteFile(outputPath, buf.Bytes(), 0o644); err != nil {
|
||||||
|
return fmt.Errorf("failed to write markdown file: %w", err)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetSupportedFormat returns the format name this exporter supports
|
// SupportedFormat returns "markdown".
|
||||||
// It indicates the file format that the MarkdownExporter can generate.
|
func (e *MarkdownExporter) SupportedFormat() string {
|
||||||
//
|
return FormatMarkdown
|
||||||
// Returns:
|
|
||||||
// - A string representing the supported format ("markdown")
|
|
||||||
func (e *MarkdownExporter) GetSupportedFormat() string {
|
|
||||||
return "markdown"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// processItemToMarkdown converts a course item into Markdown format
|
// processItemToMarkdown converts a course item into Markdown format.
|
||||||
// and appends it to the provided buffer. It handles different item types
|
// The level parameter determines the heading level (number of # characters).
|
||||||
// with appropriate Markdown formatting.
|
|
||||||
//
|
|
||||||
// Parameters:
|
|
||||||
// - buf: The buffer to write the Markdown content to
|
|
||||||
// - item: The course item to process
|
|
||||||
// - level: The heading level for the item (determines the number of # characters)
|
|
||||||
func (e *MarkdownExporter) processItemToMarkdown(buf *bytes.Buffer, item models.Item, level int) {
|
func (e *MarkdownExporter) processItemToMarkdown(buf *bytes.Buffer, item models.Item, level int) {
|
||||||
headingPrefix := strings.Repeat("#", level)
|
headingPrefix := strings.Repeat("#", level)
|
||||||
|
|
||||||
switch item.Type {
|
// Normalize item type to lowercase for consistent matching
|
||||||
|
itemType := strings.ToLower(item.Type)
|
||||||
|
|
||||||
|
switch itemType {
|
||||||
case "text":
|
case "text":
|
||||||
e.processTextItem(buf, item, headingPrefix)
|
e.processTextItem(buf, item, headingPrefix)
|
||||||
case "list":
|
case "list":
|
||||||
@ -119,7 +108,7 @@ func (e *MarkdownExporter) processItemToMarkdown(buf *bytes.Buffer, item models.
|
|||||||
e.processMultimediaItem(buf, item, headingPrefix)
|
e.processMultimediaItem(buf, item, headingPrefix)
|
||||||
case "image":
|
case "image":
|
||||||
e.processImageItem(buf, item, headingPrefix)
|
e.processImageItem(buf, item, headingPrefix)
|
||||||
case "knowledgeCheck":
|
case "knowledgecheck":
|
||||||
e.processKnowledgeCheckItem(buf, item, headingPrefix)
|
e.processKnowledgeCheckItem(buf, item, headingPrefix)
|
||||||
case "interactive":
|
case "interactive":
|
||||||
e.processInteractiveItem(buf, item, headingPrefix)
|
e.processInteractiveItem(buf, item, headingPrefix)
|
||||||
@ -130,47 +119,47 @@ func (e *MarkdownExporter) processItemToMarkdown(buf *bytes.Buffer, item models.
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processTextItem handles text content with headings and paragraphs
|
// processTextItem handles text content with headings and paragraphs.
|
||||||
func (e *MarkdownExporter) processTextItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
func (e *MarkdownExporter) processTextItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
||||||
for _, subItem := range item.Items {
|
for _, subItem := range item.Items {
|
||||||
if subItem.Heading != "" {
|
if subItem.Heading != "" {
|
||||||
heading := e.htmlCleaner.CleanHTML(subItem.Heading)
|
heading := e.htmlCleaner.CleanHTML(subItem.Heading)
|
||||||
if heading != "" {
|
if heading != "" {
|
||||||
buf.WriteString(fmt.Sprintf("%s %s\n\n", headingPrefix, heading))
|
fmt.Fprintf(buf, "%s %s\n\n", headingPrefix, heading)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if subItem.Paragraph != "" {
|
if subItem.Paragraph != "" {
|
||||||
paragraph := e.htmlCleaner.CleanHTML(subItem.Paragraph)
|
paragraph := e.htmlCleaner.CleanHTML(subItem.Paragraph)
|
||||||
if paragraph != "" {
|
if paragraph != "" {
|
||||||
buf.WriteString(fmt.Sprintf("%s\n\n", paragraph))
|
fmt.Fprintf(buf, "%s\n\n", paragraph)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processListItem handles list items with bullet points
|
// processListItem handles list items with bullet points.
|
||||||
func (e *MarkdownExporter) processListItem(buf *bytes.Buffer, item models.Item) {
|
func (e *MarkdownExporter) processListItem(buf *bytes.Buffer, item models.Item) {
|
||||||
for _, subItem := range item.Items {
|
for _, subItem := range item.Items {
|
||||||
if subItem.Paragraph != "" {
|
if subItem.Paragraph != "" {
|
||||||
paragraph := e.htmlCleaner.CleanHTML(subItem.Paragraph)
|
paragraph := e.htmlCleaner.CleanHTML(subItem.Paragraph)
|
||||||
if paragraph != "" {
|
if paragraph != "" {
|
||||||
buf.WriteString(fmt.Sprintf("- %s\n", paragraph))
|
fmt.Fprintf(buf, "- %s\n", paragraph)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
buf.WriteString("\n")
|
buf.WriteString("\n")
|
||||||
}
|
}
|
||||||
|
|
||||||
// processMultimediaItem handles multimedia content including videos and images
|
// processMultimediaItem handles multimedia content including videos and images.
|
||||||
func (e *MarkdownExporter) processMultimediaItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
func (e *MarkdownExporter) processMultimediaItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
||||||
buf.WriteString(fmt.Sprintf("%s Media Content\n\n", headingPrefix))
|
fmt.Fprintf(buf, "%s Media Content\n\n", headingPrefix)
|
||||||
for _, subItem := range item.Items {
|
for _, subItem := range item.Items {
|
||||||
e.processMediaSubItem(buf, subItem)
|
e.processMediaSubItem(buf, subItem)
|
||||||
}
|
}
|
||||||
buf.WriteString("\n")
|
buf.WriteString("\n")
|
||||||
}
|
}
|
||||||
|
|
||||||
// processMediaSubItem processes individual media items (video/image)
|
// processMediaSubItem processes individual media items (video/image).
|
||||||
func (e *MarkdownExporter) processMediaSubItem(buf *bytes.Buffer, subItem models.SubItem) {
|
func (e *MarkdownExporter) processMediaSubItem(buf *bytes.Buffer, subItem models.SubItem) {
|
||||||
if subItem.Media != nil {
|
if subItem.Media != nil {
|
||||||
e.processVideoMedia(buf, subItem.Media)
|
e.processVideoMedia(buf, subItem.Media)
|
||||||
@ -178,67 +167,67 @@ func (e *MarkdownExporter) processMediaSubItem(buf *bytes.Buffer, subItem models
|
|||||||
}
|
}
|
||||||
if subItem.Caption != "" {
|
if subItem.Caption != "" {
|
||||||
caption := e.htmlCleaner.CleanHTML(subItem.Caption)
|
caption := e.htmlCleaner.CleanHTML(subItem.Caption)
|
||||||
buf.WriteString(fmt.Sprintf("*%s*\n", caption))
|
fmt.Fprintf(buf, "*%s*\n", caption)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processVideoMedia processes video media content
|
// processVideoMedia processes video media content.
|
||||||
func (e *MarkdownExporter) processVideoMedia(buf *bytes.Buffer, media *models.Media) {
|
func (e *MarkdownExporter) processVideoMedia(buf *bytes.Buffer, media *models.Media) {
|
||||||
if media.Video != nil {
|
if media.Video != nil {
|
||||||
buf.WriteString(fmt.Sprintf("**Video**: %s\n", media.Video.OriginalUrl))
|
fmt.Fprintf(buf, "**Video**: %s\n", media.Video.OriginalURL)
|
||||||
if media.Video.Duration > 0 {
|
if media.Video.Duration > 0 {
|
||||||
buf.WriteString(fmt.Sprintf("**Duration**: %d seconds\n", media.Video.Duration))
|
fmt.Fprintf(buf, "**Duration**: %d seconds\n", media.Video.Duration)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processImageMedia processes image media content
|
// processImageMedia processes image media content.
|
||||||
func (e *MarkdownExporter) processImageMedia(buf *bytes.Buffer, media *models.Media) {
|
func (e *MarkdownExporter) processImageMedia(buf *bytes.Buffer, media *models.Media) {
|
||||||
if media.Image != nil {
|
if media.Image != nil {
|
||||||
buf.WriteString(fmt.Sprintf("**Image**: %s\n", media.Image.OriginalUrl))
|
fmt.Fprintf(buf, "**Image**: %s\n", media.Image.OriginalURL)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processImageItem handles standalone image items
|
// processImageItem handles standalone image items.
|
||||||
func (e *MarkdownExporter) processImageItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
func (e *MarkdownExporter) processImageItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
||||||
buf.WriteString(fmt.Sprintf("%s Image\n\n", headingPrefix))
|
fmt.Fprintf(buf, "%s Image\n\n", headingPrefix)
|
||||||
for _, subItem := range item.Items {
|
for _, subItem := range item.Items {
|
||||||
if subItem.Media != nil && subItem.Media.Image != nil {
|
if subItem.Media != nil && subItem.Media.Image != nil {
|
||||||
buf.WriteString(fmt.Sprintf("**Image**: %s\n", subItem.Media.Image.OriginalUrl))
|
fmt.Fprintf(buf, "**Image**: %s\n", subItem.Media.Image.OriginalURL)
|
||||||
}
|
}
|
||||||
if subItem.Caption != "" {
|
if subItem.Caption != "" {
|
||||||
caption := e.htmlCleaner.CleanHTML(subItem.Caption)
|
caption := e.htmlCleaner.CleanHTML(subItem.Caption)
|
||||||
buf.WriteString(fmt.Sprintf("*%s*\n", caption))
|
fmt.Fprintf(buf, "*%s*\n", caption)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
buf.WriteString("\n")
|
buf.WriteString("\n")
|
||||||
}
|
}
|
||||||
|
|
||||||
// processKnowledgeCheckItem handles quiz questions and knowledge checks
|
// processKnowledgeCheckItem handles quiz questions and knowledge checks.
|
||||||
func (e *MarkdownExporter) processKnowledgeCheckItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
func (e *MarkdownExporter) processKnowledgeCheckItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
||||||
buf.WriteString(fmt.Sprintf("%s Knowledge Check\n\n", headingPrefix))
|
fmt.Fprintf(buf, "%s Knowledge Check\n\n", headingPrefix)
|
||||||
for _, subItem := range item.Items {
|
for _, subItem := range item.Items {
|
||||||
e.processQuestionSubItem(buf, subItem)
|
e.processQuestionSubItem(buf, subItem)
|
||||||
}
|
}
|
||||||
buf.WriteString("\n")
|
buf.WriteString("\n")
|
||||||
}
|
}
|
||||||
|
|
||||||
// processQuestionSubItem processes individual question items
|
// processQuestionSubItem processes individual question items.
|
||||||
func (e *MarkdownExporter) processQuestionSubItem(buf *bytes.Buffer, subItem models.SubItem) {
|
func (e *MarkdownExporter) processQuestionSubItem(buf *bytes.Buffer, subItem models.SubItem) {
|
||||||
if subItem.Title != "" {
|
if subItem.Title != "" {
|
||||||
title := e.htmlCleaner.CleanHTML(subItem.Title)
|
title := e.htmlCleaner.CleanHTML(subItem.Title)
|
||||||
buf.WriteString(fmt.Sprintf("**Question**: %s\n\n", title))
|
fmt.Fprintf(buf, "**Question**: %s\n\n", title)
|
||||||
}
|
}
|
||||||
|
|
||||||
e.processAnswers(buf, subItem.Answers)
|
e.processAnswers(buf, subItem.Answers)
|
||||||
|
|
||||||
if subItem.Feedback != "" {
|
if subItem.Feedback != "" {
|
||||||
feedback := e.htmlCleaner.CleanHTML(subItem.Feedback)
|
feedback := e.htmlCleaner.CleanHTML(subItem.Feedback)
|
||||||
buf.WriteString(fmt.Sprintf("\n**Feedback**: %s\n", feedback))
|
fmt.Fprintf(buf, "\n**Feedback**: %s\n", feedback)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processAnswers processes answer choices for quiz questions
|
// processAnswers processes answer choices for quiz questions.
|
||||||
func (e *MarkdownExporter) processAnswers(buf *bytes.Buffer, answers []models.Answer) {
|
func (e *MarkdownExporter) processAnswers(buf *bytes.Buffer, answers []models.Answer) {
|
||||||
buf.WriteString("**Answers**:\n")
|
buf.WriteString("**Answers**:\n")
|
||||||
for i, answer := range answers {
|
for i, answer := range answers {
|
||||||
@ -246,44 +235,45 @@ func (e *MarkdownExporter) processAnswers(buf *bytes.Buffer, answers []models.An
|
|||||||
if answer.Correct {
|
if answer.Correct {
|
||||||
correctMark = " ✓"
|
correctMark = " ✓"
|
||||||
}
|
}
|
||||||
buf.WriteString(fmt.Sprintf("%d. %s%s\n", i+1, answer.Title, correctMark))
|
fmt.Fprintf(buf, "%d. %s%s\n", i+1, answer.Title, correctMark)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processInteractiveItem handles interactive content
|
// processInteractiveItem handles interactive content.
|
||||||
func (e *MarkdownExporter) processInteractiveItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
func (e *MarkdownExporter) processInteractiveItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
||||||
buf.WriteString(fmt.Sprintf("%s Interactive Content\n\n", headingPrefix))
|
fmt.Fprintf(buf, "%s Interactive Content\n\n", headingPrefix)
|
||||||
for _, subItem := range item.Items {
|
for _, subItem := range item.Items {
|
||||||
if subItem.Title != "" {
|
if subItem.Title != "" {
|
||||||
title := e.htmlCleaner.CleanHTML(subItem.Title)
|
title := e.htmlCleaner.CleanHTML(subItem.Title)
|
||||||
buf.WriteString(fmt.Sprintf("**%s**\n\n", title))
|
fmt.Fprintf(buf, "**%s**\n\n", title)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processDividerItem handles divider elements
|
// processDividerItem handles divider elements.
|
||||||
func (e *MarkdownExporter) processDividerItem(buf *bytes.Buffer) {
|
func (e *MarkdownExporter) processDividerItem(buf *bytes.Buffer) {
|
||||||
buf.WriteString("---\n\n")
|
buf.WriteString("---\n\n")
|
||||||
}
|
}
|
||||||
|
|
||||||
// processUnknownItem handles unknown or unsupported item types
|
// processUnknownItem handles unknown or unsupported item types.
|
||||||
func (e *MarkdownExporter) processUnknownItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
func (e *MarkdownExporter) processUnknownItem(buf *bytes.Buffer, item models.Item, headingPrefix string) {
|
||||||
if len(item.Items) > 0 {
|
if len(item.Items) > 0 {
|
||||||
buf.WriteString(fmt.Sprintf("%s %s Content\n\n", headingPrefix, strings.Title(item.Type)))
|
caser := cases.Title(language.English)
|
||||||
|
fmt.Fprintf(buf, "%s %s Content\n\n", headingPrefix, caser.String(item.Type))
|
||||||
for _, subItem := range item.Items {
|
for _, subItem := range item.Items {
|
||||||
e.processGenericSubItem(buf, subItem)
|
e.processGenericSubItem(buf, subItem)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// processGenericSubItem processes sub-items for unknown types
|
// processGenericSubItem processes sub-items for unknown types.
|
||||||
func (e *MarkdownExporter) processGenericSubItem(buf *bytes.Buffer, subItem models.SubItem) {
|
func (e *MarkdownExporter) processGenericSubItem(buf *bytes.Buffer, subItem models.SubItem) {
|
||||||
if subItem.Title != "" {
|
if subItem.Title != "" {
|
||||||
title := e.htmlCleaner.CleanHTML(subItem.Title)
|
title := e.htmlCleaner.CleanHTML(subItem.Title)
|
||||||
buf.WriteString(fmt.Sprintf("**%s**\n\n", title))
|
fmt.Fprintf(buf, "**%s**\n\n", title)
|
||||||
}
|
}
|
||||||
if subItem.Paragraph != "" {
|
if subItem.Paragraph != "" {
|
||||||
paragraph := e.htmlCleaner.CleanHTML(subItem.Paragraph)
|
paragraph := e.htmlCleaner.CleanHTML(subItem.Paragraph)
|
||||||
buf.WriteString(fmt.Sprintf("%s\n\n", paragraph))
|
fmt.Fprintf(buf, "%s\n\n", paragraph)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,4 +1,3 @@
|
|||||||
// Package exporters_test provides tests for the markdown exporter.
|
|
||||||
package exporters
|
package exporters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@ -32,13 +31,13 @@ func TestNewMarkdownExporter(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestMarkdownExporter_GetSupportedFormat tests the GetSupportedFormat method.
|
// TestMarkdownExporter_SupportedFormat tests the SupportedFormat method.
|
||||||
func TestMarkdownExporter_GetSupportedFormat(t *testing.T) {
|
func TestMarkdownExporter_SupportedFormat(t *testing.T) {
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
exporter := NewMarkdownExporter(htmlCleaner)
|
exporter := NewMarkdownExporter(htmlCleaner)
|
||||||
|
|
||||||
expected := "markdown"
|
expected := "markdown"
|
||||||
result := exporter.GetSupportedFormat()
|
result := exporter.SupportedFormat()
|
||||||
|
|
||||||
if result != expected {
|
if result != expected {
|
||||||
t.Errorf("Expected format '%s', got '%s'", expected, result)
|
t.Errorf("Expected format '%s', got '%s'", expected, result)
|
||||||
@ -188,7 +187,7 @@ func TestMarkdownExporter_ProcessMultimediaItem(t *testing.T) {
|
|||||||
{
|
{
|
||||||
Media: &models.Media{
|
Media: &models.Media{
|
||||||
Video: &models.VideoMedia{
|
Video: &models.VideoMedia{
|
||||||
OriginalUrl: "https://example.com/video.mp4",
|
OriginalURL: "https://example.com/video.mp4",
|
||||||
Duration: 120,
|
Duration: 120,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@ -227,7 +226,7 @@ func TestMarkdownExporter_ProcessImageItem(t *testing.T) {
|
|||||||
{
|
{
|
||||||
Media: &models.Media{
|
Media: &models.Media{
|
||||||
Image: &models.ImageMedia{
|
Image: &models.ImageMedia{
|
||||||
OriginalUrl: "https://example.com/image.jpg",
|
OriginalURL: "https://example.com/image.jpg",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
Caption: "<p>Image caption</p>",
|
Caption: "<p>Image caption</p>",
|
||||||
@ -372,7 +371,7 @@ func TestMarkdownExporter_ProcessVideoMedia(t *testing.T) {
|
|||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
media := &models.Media{
|
media := &models.Media{
|
||||||
Video: &models.VideoMedia{
|
Video: &models.VideoMedia{
|
||||||
OriginalUrl: "https://example.com/video.mp4",
|
OriginalURL: "https://example.com/video.mp4",
|
||||||
Duration: 300,
|
Duration: 300,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
@ -397,7 +396,7 @@ func TestMarkdownExporter_ProcessImageMedia(t *testing.T) {
|
|||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
media := &models.Media{
|
media := &models.Media{
|
||||||
Image: &models.ImageMedia{
|
Image: &models.ImageMedia{
|
||||||
OriginalUrl: "https://example.com/image.jpg",
|
OriginalURL: "https://example.com/image.jpg",
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -661,12 +660,13 @@ func BenchmarkMarkdownExporter_Export(b *testing.B) {
|
|||||||
// Create temporary directory
|
// Create temporary directory
|
||||||
tempDir := b.TempDir()
|
tempDir := b.TempDir()
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
outputPath := filepath.Join(tempDir, "benchmark-course.md")
|
outputPath := filepath.Join(tempDir, "benchmark-course.md")
|
||||||
_ = exporter.Export(course, outputPath)
|
_ = exporter.Export(course, outputPath)
|
||||||
// Clean up for next iteration
|
// Clean up for next iteration. Remove errors are ignored because we've already
|
||||||
os.Remove(outputPath)
|
// benchmarked the export operation; cleanup failures don't affect the benchmark
|
||||||
|
// measurements or the validity of the next iteration's export.
|
||||||
|
_ = os.Remove(outputPath)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -685,8 +685,7 @@ func BenchmarkMarkdownExporter_ProcessTextItem(b *testing.B) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
exporter.processTextItem(&buf, item, "###")
|
exporter.processTextItem(&buf, item, "###")
|
||||||
}
|
}
|
||||||
|
|||||||
BIN
internal/exporters/output.docx
Normal file
BIN
internal/exporters/output.docx
Normal file
Binary file not shown.
12
internal/exporters/output.md
Normal file
12
internal/exporters/output.md
Normal file
@ -0,0 +1,12 @@
|
|||||||
|
# Example Course
|
||||||
|
|
||||||
|
Course description
|
||||||
|
|
||||||
|
## Course Information
|
||||||
|
|
||||||
|
- **Course ID**:
|
||||||
|
- **Share ID**: example-id
|
||||||
|
- **Navigation Mode**:
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
@ -1,5 +1,3 @@
|
|||||||
// Package interfaces provides the core contracts for the articulate-parser application.
|
|
||||||
// It defines interfaces for parsing and exporting Articulate Rise courses.
|
|
||||||
package interfaces
|
package interfaces
|
||||||
|
|
||||||
import "github.com/kjanat/articulate-parser/internal/models"
|
import "github.com/kjanat/articulate-parser/internal/models"
|
||||||
@ -12,9 +10,9 @@ type Exporter interface {
|
|||||||
// specified output path. It returns an error if the export operation fails.
|
// specified output path. It returns an error if the export operation fails.
|
||||||
Export(course *models.Course, outputPath string) error
|
Export(course *models.Course, outputPath string) error
|
||||||
|
|
||||||
// GetSupportedFormat returns the name of the format this exporter supports.
|
// SupportedFormat returns the name of the format this exporter supports.
|
||||||
// This is used to identify which exporter to use for a given format.
|
// This is used to identify which exporter to use for a given format.
|
||||||
GetSupportedFormat() string
|
SupportedFormat() string
|
||||||
}
|
}
|
||||||
|
|
||||||
// ExporterFactory creates exporters for different formats.
|
// ExporterFactory creates exporters for different formats.
|
||||||
@ -25,7 +23,7 @@ type ExporterFactory interface {
|
|||||||
// It returns the appropriate exporter or an error if the format is not supported.
|
// It returns the appropriate exporter or an error if the format is not supported.
|
||||||
CreateExporter(format string) (Exporter, error)
|
CreateExporter(format string) (Exporter, error)
|
||||||
|
|
||||||
// GetSupportedFormats returns a list of all export formats supported by this factory.
|
// SupportedFormats returns a list of all export formats supported by this factory.
|
||||||
// This is used to inform users of available export options.
|
// This is used to inform users of available export options.
|
||||||
GetSupportedFormats() []string
|
SupportedFormats() []string
|
||||||
}
|
}
|
||||||
|
|||||||
25
internal/interfaces/logger.go
Normal file
25
internal/interfaces/logger.go
Normal file
@ -0,0 +1,25 @@
|
|||||||
|
package interfaces
|
||||||
|
|
||||||
|
import "context"
|
||||||
|
|
||||||
|
// Logger defines the interface for structured logging.
|
||||||
|
// Implementations should provide leveled, structured logging capabilities.
|
||||||
|
type Logger interface {
|
||||||
|
// Debug logs a debug-level message with optional key-value pairs.
|
||||||
|
Debug(msg string, keysAndValues ...any)
|
||||||
|
|
||||||
|
// Info logs an info-level message with optional key-value pairs.
|
||||||
|
Info(msg string, keysAndValues ...any)
|
||||||
|
|
||||||
|
// Warn logs a warning-level message with optional key-value pairs.
|
||||||
|
Warn(msg string, keysAndValues ...any)
|
||||||
|
|
||||||
|
// Error logs an error-level message with optional key-value pairs.
|
||||||
|
Error(msg string, keysAndValues ...any)
|
||||||
|
|
||||||
|
// With returns a new logger with the given key-value pairs added as context.
|
||||||
|
With(keysAndValues ...any) Logger
|
||||||
|
|
||||||
|
// WithContext returns a new logger with context information.
|
||||||
|
WithContext(ctx context.Context) Logger
|
||||||
|
}
|
||||||
@ -2,7 +2,11 @@
|
|||||||
// It defines interfaces for parsing and exporting Articulate Rise courses.
|
// It defines interfaces for parsing and exporting Articulate Rise courses.
|
||||||
package interfaces
|
package interfaces
|
||||||
|
|
||||||
import "github.com/kjanat/articulate-parser/internal/models"
|
import (
|
||||||
|
"context"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
|
)
|
||||||
|
|
||||||
// CourseParser defines the interface for loading course data.
|
// CourseParser defines the interface for loading course data.
|
||||||
// It provides methods to fetch course content either from a remote URI
|
// It provides methods to fetch course content either from a remote URI
|
||||||
@ -10,8 +14,9 @@ import "github.com/kjanat/articulate-parser/internal/models"
|
|||||||
type CourseParser interface {
|
type CourseParser interface {
|
||||||
// FetchCourse loads a course from a URI (typically an Articulate Rise share URL).
|
// FetchCourse loads a course from a URI (typically an Articulate Rise share URL).
|
||||||
// It retrieves the course data from the remote location and returns a parsed Course model.
|
// It retrieves the course data from the remote location and returns a parsed Course model.
|
||||||
|
// The context can be used for cancellation and timeout control.
|
||||||
// Returns an error if the fetch operation fails or if the data cannot be parsed.
|
// Returns an error if the fetch operation fails or if the data cannot be parsed.
|
||||||
FetchCourse(uri string) (*models.Course, error)
|
FetchCourse(ctx context.Context, uri string) (*models.Course, error)
|
||||||
|
|
||||||
// LoadCourseFromFile loads a course from a local file.
|
// LoadCourseFromFile loads a course from a local file.
|
||||||
// It reads and parses the course data from the specified file path.
|
// It reads and parses the course data from the specified file path.
|
||||||
|
|||||||
@ -1,5 +1,3 @@
|
|||||||
// Package models defines the data structures representing Articulate Rise courses.
|
|
||||||
// These structures closely match the JSON format used by Articulate Rise.
|
|
||||||
package models
|
package models
|
||||||
|
|
||||||
// Lesson represents a single lesson or section within an Articulate Rise course.
|
// Lesson represents a single lesson or section within an Articulate Rise course.
|
||||||
@ -18,7 +16,7 @@ type Lesson struct {
|
|||||||
// Items is an ordered array of content items within the lesson
|
// Items is an ordered array of content items within the lesson
|
||||||
Items []Item `json:"items"`
|
Items []Item `json:"items"`
|
||||||
// Position stores the ordering information for the lesson
|
// Position stores the ordering information for the lesson
|
||||||
Position interface{} `json:"position"`
|
Position any `json:"position"`
|
||||||
// Ready indicates whether the lesson is marked as complete
|
// Ready indicates whether the lesson is marked as complete
|
||||||
Ready bool `json:"ready"`
|
Ready bool `json:"ready"`
|
||||||
// CreatedAt is the timestamp when the lesson was created
|
// CreatedAt is the timestamp when the lesson was created
|
||||||
@ -41,9 +39,9 @@ type Item struct {
|
|||||||
// Items contains the actual content elements (sub-items) of this item
|
// Items contains the actual content elements (sub-items) of this item
|
||||||
Items []SubItem `json:"items"`
|
Items []SubItem `json:"items"`
|
||||||
// Settings contains configuration options specific to this item type
|
// Settings contains configuration options specific to this item type
|
||||||
Settings interface{} `json:"settings"`
|
Settings any `json:"settings"`
|
||||||
// Data contains additional structured data for the item
|
// Data contains additional structured data for the item
|
||||||
Data interface{} `json:"data"`
|
Data any `json:"data"`
|
||||||
// Media contains any associated media for the item
|
// Media contains any associated media for the item
|
||||||
Media *Media `json:"media,omitempty"`
|
Media *Media `json:"media,omitempty"`
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,5 +1,3 @@
|
|||||||
// Package models defines the data structures representing Articulate Rise courses.
|
|
||||||
// These structures closely match the JSON format used by Articulate Rise.
|
|
||||||
package models
|
package models
|
||||||
|
|
||||||
// Media represents a media element that can be either an image or a video.
|
// Media represents a media element that can be either an image or a video.
|
||||||
@ -23,8 +21,8 @@ type ImageMedia struct {
|
|||||||
Height int `json:"height,omitempty"`
|
Height int `json:"height,omitempty"`
|
||||||
// CrushedKey is the identifier for a compressed version of the image
|
// CrushedKey is the identifier for a compressed version of the image
|
||||||
CrushedKey string `json:"crushedKey,omitempty"`
|
CrushedKey string `json:"crushedKey,omitempty"`
|
||||||
// OriginalUrl is the URL to the full-resolution image
|
// OriginalURL is the URL to the full-resolution image
|
||||||
OriginalUrl string `json:"originalUrl"`
|
OriginalURL string `json:"originalUrl"`
|
||||||
// UseCrushedKey indicates whether to use the compressed version
|
// UseCrushedKey indicates whether to use the compressed version
|
||||||
UseCrushedKey bool `json:"useCrushedKey,omitempty"`
|
UseCrushedKey bool `json:"useCrushedKey,omitempty"`
|
||||||
}
|
}
|
||||||
@ -45,6 +43,6 @@ type VideoMedia struct {
|
|||||||
InputKey string `json:"inputKey,omitempty"`
|
InputKey string `json:"inputKey,omitempty"`
|
||||||
// Thumbnail is the URL to a smaller preview image
|
// Thumbnail is the URL to a smaller preview image
|
||||||
Thumbnail string `json:"thumbnail,omitempty"`
|
Thumbnail string `json:"thumbnail,omitempty"`
|
||||||
// OriginalUrl is the URL to the source video file
|
// OriginalURL is the URL to the source video file
|
||||||
OriginalUrl string `json:"originalUrl"`
|
OriginalURL string `json:"originalUrl"`
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,4 +1,3 @@
|
|||||||
// Package models_test provides tests for the data models.
|
|
||||||
package models
|
package models
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@ -98,7 +97,7 @@ func TestCourseInfo_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
Type: "jpg",
|
Type: "jpg",
|
||||||
Width: 800,
|
Width: 800,
|
||||||
Height: 600,
|
Height: 600,
|
||||||
OriginalUrl: "https://example.com/image.jpg",
|
OriginalURL: "https://example.com/image.jpg",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
@ -133,7 +132,7 @@ func TestLesson_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
Ready: true,
|
Ready: true,
|
||||||
CreatedAt: "2023-06-01T12:00:00Z",
|
CreatedAt: "2023-06-01T12:00:00Z",
|
||||||
UpdatedAt: "2023-06-01T13:00:00Z",
|
UpdatedAt: "2023-06-01T13:00:00Z",
|
||||||
Position: map[string]interface{}{"x": 1, "y": 2},
|
Position: map[string]any{"x": 1, "y": 2},
|
||||||
Items: []Item{
|
Items: []Item{
|
||||||
{
|
{
|
||||||
ID: "item-test",
|
ID: "item-test",
|
||||||
@ -149,13 +148,13 @@ func TestLesson_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
URL: "https://example.com/video.mp4",
|
URL: "https://example.com/video.mp4",
|
||||||
Type: "mp4",
|
Type: "mp4",
|
||||||
Duration: 120,
|
Duration: 120,
|
||||||
OriginalUrl: "https://example.com/video.mp4",
|
OriginalURL: "https://example.com/video.mp4",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
Settings: map[string]interface{}{"autoplay": false},
|
Settings: map[string]any{"autoplay": false},
|
||||||
Data: map[string]interface{}{"metadata": "test"},
|
Data: map[string]any{"metadata": "test"},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
@ -197,11 +196,11 @@ func TestItem_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
Feedback: "Well done!",
|
Feedback: "Well done!",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
Settings: map[string]interface{}{
|
Settings: map[string]any{
|
||||||
"allowRetry": true,
|
"allowRetry": true,
|
||||||
"showAnswer": true,
|
"showAnswer": true,
|
||||||
},
|
},
|
||||||
Data: map[string]interface{}{
|
Data: map[string]any{
|
||||||
"points": 10,
|
"points": 10,
|
||||||
"weight": 1.5,
|
"weight": 1.5,
|
||||||
},
|
},
|
||||||
@ -244,7 +243,7 @@ func TestSubItem_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
Type: "png",
|
Type: "png",
|
||||||
Width: 400,
|
Width: 400,
|
||||||
Height: 300,
|
Height: 300,
|
||||||
OriginalUrl: "https://example.com/subitem.png",
|
OriginalURL: "https://example.com/subitem.png",
|
||||||
CrushedKey: "crushed-123",
|
CrushedKey: "crushed-123",
|
||||||
UseCrushedKey: true,
|
UseCrushedKey: true,
|
||||||
},
|
},
|
||||||
@ -305,7 +304,7 @@ func TestMedia_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
Type: "jpeg",
|
Type: "jpeg",
|
||||||
Width: 1200,
|
Width: 1200,
|
||||||
Height: 800,
|
Height: 800,
|
||||||
OriginalUrl: "https://example.com/media.jpg",
|
OriginalURL: "https://example.com/media.jpg",
|
||||||
CrushedKey: "crushed-media",
|
CrushedKey: "crushed-media",
|
||||||
UseCrushedKey: false,
|
UseCrushedKey: false,
|
||||||
},
|
},
|
||||||
@ -336,7 +335,7 @@ func TestMedia_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
Poster: "https://example.com/poster.jpg",
|
Poster: "https://example.com/poster.jpg",
|
||||||
Thumbnail: "https://example.com/thumb.jpg",
|
Thumbnail: "https://example.com/thumb.jpg",
|
||||||
InputKey: "input-123",
|
InputKey: "input-123",
|
||||||
OriginalUrl: "https://example.com/original.mp4",
|
OriginalURL: "https://example.com/original.mp4",
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -363,7 +362,7 @@ func TestImageMedia_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
Type: "gif",
|
Type: "gif",
|
||||||
Width: 640,
|
Width: 640,
|
||||||
Height: 480,
|
Height: 480,
|
||||||
OriginalUrl: "https://example.com/image.gif",
|
OriginalURL: "https://example.com/image.gif",
|
||||||
CrushedKey: "crushed-gif",
|
CrushedKey: "crushed-gif",
|
||||||
UseCrushedKey: true,
|
UseCrushedKey: true,
|
||||||
}
|
}
|
||||||
@ -397,7 +396,7 @@ func TestVideoMedia_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
Poster: "https://example.com/poster.jpg",
|
Poster: "https://example.com/poster.jpg",
|
||||||
Thumbnail: "https://example.com/thumbnail.jpg",
|
Thumbnail: "https://example.com/thumbnail.jpg",
|
||||||
InputKey: "upload-456",
|
InputKey: "upload-456",
|
||||||
OriginalUrl: "https://example.com/original.webm",
|
OriginalURL: "https://example.com/original.webm",
|
||||||
}
|
}
|
||||||
|
|
||||||
// Marshal to JSON
|
// Marshal to JSON
|
||||||
@ -475,7 +474,7 @@ func TestLabelSet_JSONMarshalUnmarshal(t *testing.T) {
|
|||||||
func TestEmptyStructures(t *testing.T) {
|
func TestEmptyStructures(t *testing.T) {
|
||||||
testCases := []struct {
|
testCases := []struct {
|
||||||
name string
|
name string
|
||||||
data interface{}
|
data any
|
||||||
}{
|
}{
|
||||||
{"Empty Course", Course{}},
|
{"Empty Course", Course{}},
|
||||||
{"Empty CourseInfo", CourseInfo{}},
|
{"Empty CourseInfo", CourseInfo{}},
|
||||||
@ -569,7 +568,7 @@ func TestNilPointerSafety(t *testing.T) {
|
|||||||
// TestJSONTagsPresence tests that JSON tags are properly defined.
|
// TestJSONTagsPresence tests that JSON tags are properly defined.
|
||||||
func TestJSONTagsPresence(t *testing.T) {
|
func TestJSONTagsPresence(t *testing.T) {
|
||||||
// Test that important fields have JSON tags
|
// Test that important fields have JSON tags
|
||||||
courseType := reflect.TypeOf(Course{})
|
courseType := reflect.TypeFor[Course]()
|
||||||
if courseType.Kind() == reflect.Struct {
|
if courseType.Kind() == reflect.Struct {
|
||||||
field, found := courseType.FieldByName("ShareID")
|
field, found := courseType.FieldByName("ShareID")
|
||||||
if !found {
|
if !found {
|
||||||
@ -586,7 +585,7 @@ func TestJSONTagsPresence(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Test CourseInfo
|
// Test CourseInfo
|
||||||
courseInfoType := reflect.TypeOf(CourseInfo{})
|
courseInfoType := reflect.TypeFor[CourseInfo]()
|
||||||
if courseInfoType.Kind() == reflect.Struct {
|
if courseInfoType.Kind() == reflect.Struct {
|
||||||
field, found := courseInfoType.FieldByName("NavigationMode")
|
field, found := courseInfoType.FieldByName("NavigationMode")
|
||||||
if !found {
|
if !found {
|
||||||
@ -626,8 +625,7 @@ func BenchmarkCourse_JSONMarshal(b *testing.B) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
_, _ = json.Marshal(course)
|
_, _ = json.Marshal(course)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -660,17 +658,16 @@ func BenchmarkCourse_JSONUnmarshal(b *testing.B) {
|
|||||||
|
|
||||||
jsonData, _ := json.Marshal(course)
|
jsonData, _ := json.Marshal(course)
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
var result Course
|
var result Course
|
||||||
_ = json.Unmarshal(jsonData, &result)
|
_ = json.Unmarshal(jsonData, &result)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// compareMaps compares two interface{} values that should be maps
|
// compareMaps compares two any values that should be maps.
|
||||||
func compareMaps(original, unmarshaled interface{}) bool {
|
func compareMaps(original, unmarshaled any) bool {
|
||||||
origMap, origOk := original.(map[string]interface{})
|
origMap, origOk := original.(map[string]any)
|
||||||
unMap, unOk := unmarshaled.(map[string]interface{})
|
unMap, unOk := unmarshaled.(map[string]any)
|
||||||
|
|
||||||
if !origOk || !unOk {
|
if !origOk || !unOk {
|
||||||
// If not maps, use deep equal
|
// If not maps, use deep equal
|
||||||
@ -714,7 +711,7 @@ func compareMaps(original, unmarshaled interface{}) bool {
|
|||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
// compareLessons compares two Lesson structs accounting for JSON type conversion
|
// compareLessons compares two Lesson structs accounting for JSON type conversion.
|
||||||
func compareLessons(original, unmarshaled Lesson) bool {
|
func compareLessons(original, unmarshaled Lesson) bool {
|
||||||
// Compare all fields except Position and Items
|
// Compare all fields except Position and Items
|
||||||
if original.ID != unmarshaled.ID ||
|
if original.ID != unmarshaled.ID ||
|
||||||
@ -737,7 +734,7 @@ func compareLessons(original, unmarshaled Lesson) bool {
|
|||||||
return compareItems(original.Items, unmarshaled.Items)
|
return compareItems(original.Items, unmarshaled.Items)
|
||||||
}
|
}
|
||||||
|
|
||||||
// compareItems compares two Item slices accounting for JSON type conversion
|
// compareItems compares two Item slices accounting for JSON type conversion.
|
||||||
func compareItems(original, unmarshaled []Item) bool {
|
func compareItems(original, unmarshaled []Item) bool {
|
||||||
if len(original) != len(unmarshaled) {
|
if len(original) != len(unmarshaled) {
|
||||||
return false
|
return false
|
||||||
@ -751,7 +748,7 @@ func compareItems(original, unmarshaled []Item) bool {
|
|||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
// compareItem compares two Item structs accounting for JSON type conversion
|
// compareItem compares two Item structs accounting for JSON type conversion.
|
||||||
func compareItem(original, unmarshaled Item) bool {
|
func compareItem(original, unmarshaled Item) bool {
|
||||||
// Compare basic fields
|
// Compare basic fields
|
||||||
if original.ID != unmarshaled.ID ||
|
if original.ID != unmarshaled.ID ||
|
||||||
|
|||||||
@ -3,6 +3,7 @@
|
|||||||
package services
|
package services
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
"github.com/kjanat/articulate-parser/internal/interfaces"
|
"github.com/kjanat/articulate-parser/internal/interfaces"
|
||||||
@ -44,8 +45,8 @@ func (a *App) ProcessCourseFromFile(filePath, format, outputPath string) error {
|
|||||||
// ProcessCourseFromURI fetches a course from the provided URI and exports it to the specified format.
|
// ProcessCourseFromURI fetches a course from the provided URI and exports it to the specified format.
|
||||||
// It takes the URI to fetch the course from, the desired export format, and the output file path.
|
// It takes the URI to fetch the course from, the desired export format, and the output file path.
|
||||||
// Returns an error if fetching or exporting fails.
|
// Returns an error if fetching or exporting fails.
|
||||||
func (a *App) ProcessCourseFromURI(uri, format, outputPath string) error {
|
func (a *App) ProcessCourseFromURI(ctx context.Context, uri, format, outputPath string) error {
|
||||||
course, err := a.parser.FetchCourse(uri)
|
course, err := a.parser.FetchCourse(ctx, uri)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to fetch course: %w", err)
|
return fmt.Errorf("failed to fetch course: %w", err)
|
||||||
}
|
}
|
||||||
@ -69,8 +70,8 @@ func (a *App) exportCourse(course *models.Course, format, outputPath string) err
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetSupportedFormats returns a list of all export formats supported by the application.
|
// SupportedFormats returns a list of all export formats supported by the application.
|
||||||
// This information is provided by the ExporterFactory.
|
// This information is provided by the ExporterFactory.
|
||||||
func (a *App) GetSupportedFormats() []string {
|
func (a *App) SupportedFormats() []string {
|
||||||
return a.exporterFactory.GetSupportedFormats()
|
return a.exporterFactory.SupportedFormats()
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,7 +1,7 @@
|
|||||||
// Package services_test provides tests for the services package.
|
|
||||||
package services
|
package services
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"errors"
|
"errors"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
@ -11,13 +11,13 @@ import (
|
|||||||
|
|
||||||
// MockCourseParser is a mock implementation of interfaces.CourseParser for testing.
|
// MockCourseParser is a mock implementation of interfaces.CourseParser for testing.
|
||||||
type MockCourseParser struct {
|
type MockCourseParser struct {
|
||||||
mockFetchCourse func(uri string) (*models.Course, error)
|
mockFetchCourse func(ctx context.Context, uri string) (*models.Course, error)
|
||||||
mockLoadCourseFromFile func(filePath string) (*models.Course, error)
|
mockLoadCourseFromFile func(filePath string) (*models.Course, error)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *MockCourseParser) FetchCourse(uri string) (*models.Course, error) {
|
func (m *MockCourseParser) FetchCourse(ctx context.Context, uri string) (*models.Course, error) {
|
||||||
if m.mockFetchCourse != nil {
|
if m.mockFetchCourse != nil {
|
||||||
return m.mockFetchCourse(uri)
|
return m.mockFetchCourse(ctx, uri)
|
||||||
}
|
}
|
||||||
return nil, errors.New("not implemented")
|
return nil, errors.New("not implemented")
|
||||||
}
|
}
|
||||||
@ -31,8 +31,8 @@ func (m *MockCourseParser) LoadCourseFromFile(filePath string) (*models.Course,
|
|||||||
|
|
||||||
// MockExporter is a mock implementation of interfaces.Exporter for testing.
|
// MockExporter is a mock implementation of interfaces.Exporter for testing.
|
||||||
type MockExporter struct {
|
type MockExporter struct {
|
||||||
mockExport func(course *models.Course, outputPath string) error
|
mockExport func(course *models.Course, outputPath string) error
|
||||||
mockGetSupportedFormat func() string
|
mockSupportedFormat func() string
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *MockExporter) Export(course *models.Course, outputPath string) error {
|
func (m *MockExporter) Export(course *models.Course, outputPath string) error {
|
||||||
@ -42,17 +42,17 @@ func (m *MockExporter) Export(course *models.Course, outputPath string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *MockExporter) GetSupportedFormat() string {
|
func (m *MockExporter) SupportedFormat() string {
|
||||||
if m.mockGetSupportedFormat != nil {
|
if m.mockSupportedFormat != nil {
|
||||||
return m.mockGetSupportedFormat()
|
return m.mockSupportedFormat()
|
||||||
}
|
}
|
||||||
return "mock"
|
return "mock"
|
||||||
}
|
}
|
||||||
|
|
||||||
// MockExporterFactory is a mock implementation of interfaces.ExporterFactory for testing.
|
// MockExporterFactory is a mock implementation of interfaces.ExporterFactory for testing.
|
||||||
type MockExporterFactory struct {
|
type MockExporterFactory struct {
|
||||||
mockCreateExporter func(format string) (*MockExporter, error)
|
mockCreateExporter func(format string) (*MockExporter, error)
|
||||||
mockGetSupportedFormats func() []string
|
mockSupportedFormats func() []string
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *MockExporterFactory) CreateExporter(format string) (interfaces.Exporter, error) {
|
func (m *MockExporterFactory) CreateExporter(format string) (interfaces.Exporter, error) {
|
||||||
@ -63,9 +63,9 @@ func (m *MockExporterFactory) CreateExporter(format string) (interfaces.Exporter
|
|||||||
return &MockExporter{}, nil
|
return &MockExporter{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *MockExporterFactory) GetSupportedFormats() []string {
|
func (m *MockExporterFactory) SupportedFormats() []string {
|
||||||
if m.mockGetSupportedFormats != nil {
|
if m.mockSupportedFormats != nil {
|
||||||
return m.mockGetSupportedFormats()
|
return m.mockSupportedFormats()
|
||||||
}
|
}
|
||||||
return []string{"mock"}
|
return []string{"mock"}
|
||||||
}
|
}
|
||||||
@ -119,7 +119,7 @@ func TestNewApp(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Test that the factory is set (we can't directly compare interface values)
|
// Test that the factory is set (we can't directly compare interface values)
|
||||||
formats := app.GetSupportedFormats()
|
formats := app.SupportedFormats()
|
||||||
if len(formats) == 0 {
|
if len(formats) == 0 {
|
||||||
t.Error("App exporterFactory was not set correctly - no supported formats")
|
t.Error("App exporterFactory was not set correctly - no supported formats")
|
||||||
}
|
}
|
||||||
@ -216,10 +216,8 @@ func TestApp_ProcessCourseFromFile(t *testing.T) {
|
|||||||
if !contains(err.Error(), tt.expectedError) {
|
if !contains(err.Error(), tt.expectedError) {
|
||||||
t.Errorf("Expected error containing '%s', got '%s'", tt.expectedError, err.Error())
|
t.Errorf("Expected error containing '%s', got '%s'", tt.expectedError, err.Error())
|
||||||
}
|
}
|
||||||
} else {
|
} else if err != nil {
|
||||||
if err != nil {
|
t.Errorf("Expected no error, got: %v", err)
|
||||||
t.Errorf("Expected no error, got: %v", err)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@ -243,7 +241,7 @@ func TestApp_ProcessCourseFromURI(t *testing.T) {
|
|||||||
format: "docx",
|
format: "docx",
|
||||||
outputPath: "output.docx",
|
outputPath: "output.docx",
|
||||||
setupMocks: func(parser *MockCourseParser, factory *MockExporterFactory, exporter *MockExporter) {
|
setupMocks: func(parser *MockCourseParser, factory *MockExporterFactory, exporter *MockExporter) {
|
||||||
parser.mockFetchCourse = func(uri string) (*models.Course, error) {
|
parser.mockFetchCourse = func(ctx context.Context, uri string) (*models.Course, error) {
|
||||||
if uri != "https://rise.articulate.com/share/test123" {
|
if uri != "https://rise.articulate.com/share/test123" {
|
||||||
t.Errorf("Expected uri 'https://rise.articulate.com/share/test123', got '%s'", uri)
|
t.Errorf("Expected uri 'https://rise.articulate.com/share/test123', got '%s'", uri)
|
||||||
}
|
}
|
||||||
@ -271,7 +269,7 @@ func TestApp_ProcessCourseFromURI(t *testing.T) {
|
|||||||
format: "docx",
|
format: "docx",
|
||||||
outputPath: "output.docx",
|
outputPath: "output.docx",
|
||||||
setupMocks: func(parser *MockCourseParser, factory *MockExporterFactory, exporter *MockExporter) {
|
setupMocks: func(parser *MockCourseParser, factory *MockExporterFactory, exporter *MockExporter) {
|
||||||
parser.mockFetchCourse = func(uri string) (*models.Course, error) {
|
parser.mockFetchCourse = func(ctx context.Context, uri string) (*models.Course, error) {
|
||||||
return nil, errors.New("network error")
|
return nil, errors.New("network error")
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@ -288,7 +286,7 @@ func TestApp_ProcessCourseFromURI(t *testing.T) {
|
|||||||
tt.setupMocks(parser, factory, exporter)
|
tt.setupMocks(parser, factory, exporter)
|
||||||
|
|
||||||
app := NewApp(parser, factory)
|
app := NewApp(parser, factory)
|
||||||
err := app.ProcessCourseFromURI(tt.uri, tt.format, tt.outputPath)
|
err := app.ProcessCourseFromURI(context.Background(), tt.uri, tt.format, tt.outputPath)
|
||||||
|
|
||||||
if tt.expectedError != "" {
|
if tt.expectedError != "" {
|
||||||
if err == nil {
|
if err == nil {
|
||||||
@ -297,28 +295,26 @@ func TestApp_ProcessCourseFromURI(t *testing.T) {
|
|||||||
if !contains(err.Error(), tt.expectedError) {
|
if !contains(err.Error(), tt.expectedError) {
|
||||||
t.Errorf("Expected error containing '%s', got '%s'", tt.expectedError, err.Error())
|
t.Errorf("Expected error containing '%s', got '%s'", tt.expectedError, err.Error())
|
||||||
}
|
}
|
||||||
} else {
|
} else if err != nil {
|
||||||
if err != nil {
|
t.Errorf("Expected no error, got: %v", err)
|
||||||
t.Errorf("Expected no error, got: %v", err)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestApp_GetSupportedFormats tests the GetSupportedFormats method.
|
// TestApp_SupportedFormats tests the SupportedFormats method.
|
||||||
func TestApp_GetSupportedFormats(t *testing.T) {
|
func TestApp_SupportedFormats(t *testing.T) {
|
||||||
expectedFormats := []string{"markdown", "docx", "pdf"}
|
expectedFormats := []string{"markdown", "docx", "pdf"}
|
||||||
|
|
||||||
parser := &MockCourseParser{}
|
parser := &MockCourseParser{}
|
||||||
factory := &MockExporterFactory{
|
factory := &MockExporterFactory{
|
||||||
mockGetSupportedFormats: func() []string {
|
mockSupportedFormats: func() []string {
|
||||||
return expectedFormats
|
return expectedFormats
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
app := NewApp(parser, factory)
|
app := NewApp(parser, factory)
|
||||||
formats := app.GetSupportedFormats()
|
formats := app.SupportedFormats()
|
||||||
|
|
||||||
if len(formats) != len(expectedFormats) {
|
if len(formats) != len(expectedFormats) {
|
||||||
t.Errorf("Expected %d formats, got %d", len(expectedFormats), len(formats))
|
t.Errorf("Expected %d formats, got %d", len(expectedFormats), len(formats))
|
||||||
@ -334,7 +330,7 @@ func TestApp_GetSupportedFormats(t *testing.T) {
|
|||||||
// contains checks if a string contains a substring.
|
// contains checks if a string contains a substring.
|
||||||
func contains(s, substr string) bool {
|
func contains(s, substr string) bool {
|
||||||
return len(s) >= len(substr) &&
|
return len(s) >= len(substr) &&
|
||||||
(len(substr) == 0 ||
|
(substr == "" ||
|
||||||
s == substr ||
|
s == substr ||
|
||||||
(len(s) > len(substr) &&
|
(len(s) > len(substr) &&
|
||||||
(s[:len(substr)] == substr ||
|
(s[:len(substr)] == substr ||
|
||||||
|
|||||||
96
internal/services/example_test.go
Normal file
96
internal/services/example_test.go
Normal file
@ -0,0 +1,96 @@
|
|||||||
|
// Package services_test provides examples for the services package.
|
||||||
|
package services_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ExampleNewArticulateParser demonstrates creating a new parser.
|
||||||
|
func ExampleNewArticulateParser() {
|
||||||
|
// Create a no-op logger for this example
|
||||||
|
logger := services.NewNoOpLogger()
|
||||||
|
|
||||||
|
// Create parser with defaults
|
||||||
|
parser := services.NewArticulateParser(logger, "", 0)
|
||||||
|
|
||||||
|
fmt.Printf("Parser created: %T\n", parser)
|
||||||
|
// Output: Parser created: *services.ArticulateParser
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleNewArticulateParser_custom demonstrates creating a parser with custom configuration.
|
||||||
|
func ExampleNewArticulateParser_custom() {
|
||||||
|
logger := services.NewNoOpLogger()
|
||||||
|
|
||||||
|
// Create parser with custom base URL and timeout
|
||||||
|
parser := services.NewArticulateParser(
|
||||||
|
logger,
|
||||||
|
"https://custom.articulate.com",
|
||||||
|
60_000_000_000, // 60 seconds in nanoseconds
|
||||||
|
)
|
||||||
|
|
||||||
|
fmt.Printf("Parser configured: %T\n", parser)
|
||||||
|
// Output: Parser configured: *services.ArticulateParser
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleArticulateParser_LoadCourseFromFile demonstrates loading a course from a file.
|
||||||
|
func ExampleArticulateParser_LoadCourseFromFile() {
|
||||||
|
logger := services.NewNoOpLogger()
|
||||||
|
parser := services.NewArticulateParser(logger, "", 0)
|
||||||
|
|
||||||
|
// In a real scenario, you'd have an actual file
|
||||||
|
// This example shows the API usage
|
||||||
|
_, err := parser.LoadCourseFromFile("course.json")
|
||||||
|
if err != nil {
|
||||||
|
log.Printf("Failed to load course: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleArticulateParser_FetchCourse demonstrates fetching a course from a URI.
|
||||||
|
func ExampleArticulateParser_FetchCourse() {
|
||||||
|
logger := services.NewNoOpLogger()
|
||||||
|
parser := services.NewArticulateParser(logger, "", 0)
|
||||||
|
|
||||||
|
// Create a context with timeout
|
||||||
|
ctx := context.Background()
|
||||||
|
|
||||||
|
// In a real scenario, you'd use an actual share URL
|
||||||
|
_, err := parser.FetchCourse(ctx, "https://rise.articulate.com/share/YOUR_SHARE_ID")
|
||||||
|
if err != nil {
|
||||||
|
log.Printf("Failed to fetch course: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleHTMLCleaner demonstrates cleaning HTML content.
|
||||||
|
func ExampleHTMLCleaner() {
|
||||||
|
cleaner := services.NewHTMLCleaner()
|
||||||
|
|
||||||
|
html := "<p>This is <strong>bold</strong> text with entities.</p>"
|
||||||
|
clean := cleaner.CleanHTML(html)
|
||||||
|
|
||||||
|
fmt.Println(clean)
|
||||||
|
// Output: This is bold text with entities.
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExampleHTMLCleaner_CleanHTML demonstrates complex HTML cleaning.
|
||||||
|
func ExampleHTMLCleaner_CleanHTML() {
|
||||||
|
cleaner := services.NewHTMLCleaner()
|
||||||
|
|
||||||
|
html := `
|
||||||
|
<div>
|
||||||
|
<h1>Title</h1>
|
||||||
|
<p>Paragraph with <a href="#">link</a> and & entity.</p>
|
||||||
|
<ul>
|
||||||
|
<li>Item 1</li>
|
||||||
|
<li>Item 2</li>
|
||||||
|
</ul>
|
||||||
|
</div>
|
||||||
|
`
|
||||||
|
clean := cleaner.CleanHTML(html)
|
||||||
|
|
||||||
|
fmt.Println(clean)
|
||||||
|
// Output: Title Paragraph with link and & entity. Item 1 Item 2
|
||||||
|
}
|
||||||
@ -1,15 +1,17 @@
|
|||||||
// Package services provides the core functionality for the articulate-parser application.
|
|
||||||
// It implements the interfaces defined in the interfaces package.
|
|
||||||
package services
|
package services
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"regexp"
|
"bytes"
|
||||||
|
stdhtml "html"
|
||||||
|
"io"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"golang.org/x/net/html"
|
||||||
)
|
)
|
||||||
|
|
||||||
// HTMLCleaner provides utilities for converting HTML content to plain text.
|
// HTMLCleaner provides utilities for converting HTML content to plain text.
|
||||||
// It removes HTML tags while preserving their content and converts HTML entities
|
// It removes HTML tags while preserving their content and converts HTML entities
|
||||||
// to their plain text equivalents.
|
// to their plain text equivalents using proper HTML parsing instead of regex.
|
||||||
type HTMLCleaner struct{}
|
type HTMLCleaner struct{}
|
||||||
|
|
||||||
// NewHTMLCleaner creates a new HTML cleaner instance.
|
// NewHTMLCleaner creates a new HTML cleaner instance.
|
||||||
@ -20,34 +22,47 @@ func NewHTMLCleaner() *HTMLCleaner {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// CleanHTML removes HTML tags and converts entities, returning clean plain text.
|
// CleanHTML removes HTML tags and converts entities, returning clean plain text.
|
||||||
// The function preserves the textual content of the HTML while removing markup.
|
// It parses the HTML into a node tree and extracts only text content,
|
||||||
// It handles common HTML entities like , &, etc., and normalizes whitespace.
|
// skipping script and style tags. HTML entities are automatically handled
|
||||||
//
|
// by the parser, and whitespace is normalized.
|
||||||
// Parameters:
|
func (h *HTMLCleaner) CleanHTML(htmlStr string) string {
|
||||||
// - html: The HTML content to clean
|
// Parse the HTML into a node tree
|
||||||
//
|
doc, err := html.Parse(strings.NewReader(htmlStr))
|
||||||
// Returns:
|
if err != nil {
|
||||||
// - A plain text string with all HTML elements and entities removed/converted
|
// If parsing fails, return empty string
|
||||||
func (h *HTMLCleaner) CleanHTML(html string) string {
|
// This maintains backward compatibility with the test expectations
|
||||||
// Remove HTML tags but preserve content
|
return ""
|
||||||
re := regexp.MustCompile(`<[^>]*>`)
|
}
|
||||||
cleaned := re.ReplaceAllString(html, "")
|
|
||||||
|
|
||||||
// Replace common HTML entities with their character equivalents
|
// Extract text content from the node tree
|
||||||
cleaned = strings.ReplaceAll(cleaned, " ", " ")
|
var buf bytes.Buffer
|
||||||
cleaned = strings.ReplaceAll(cleaned, "&", "&")
|
extractText(&buf, doc)
|
||||||
cleaned = strings.ReplaceAll(cleaned, "<", "<")
|
|
||||||
cleaned = strings.ReplaceAll(cleaned, ">", ">")
|
|
||||||
cleaned = strings.ReplaceAll(cleaned, """, "\"")
|
|
||||||
cleaned = strings.ReplaceAll(cleaned, "'", "'")
|
|
||||||
cleaned = strings.ReplaceAll(cleaned, "ï", "ï")
|
|
||||||
cleaned = strings.ReplaceAll(cleaned, "ë", "ë")
|
|
||||||
cleaned = strings.ReplaceAll(cleaned, "é", "é")
|
|
||||||
|
|
||||||
// Clean up extra whitespace by replacing multiple spaces, tabs, and newlines
|
// Unescape any remaining HTML entities
|
||||||
// with a single space, then trim any leading/trailing whitespace
|
unescaped := stdhtml.UnescapeString(buf.String())
|
||||||
cleaned = regexp.MustCompile(`\s+`).ReplaceAllString(cleaned, " ")
|
|
||||||
cleaned = strings.TrimSpace(cleaned)
|
|
||||||
|
|
||||||
return cleaned
|
// Normalize whitespace: replace multiple spaces, tabs, and newlines with a single space
|
||||||
|
cleaned := strings.Join(strings.Fields(unescaped), " ")
|
||||||
|
return strings.TrimSpace(cleaned)
|
||||||
|
}
|
||||||
|
|
||||||
|
// extractText recursively traverses the HTML node tree and extracts text content.
|
||||||
|
// It skips script and style tags to avoid including their content in the output.
|
||||||
|
func extractText(w io.Writer, n *html.Node) {
|
||||||
|
// Skip script and style tags entirely
|
||||||
|
if n.Type == html.ElementNode && (n.Data == "script" || n.Data == "style") {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// If this is a text node, write its content
|
||||||
|
if n.Type == html.TextNode {
|
||||||
|
// Write errors are ignored because we're writing to an in-memory buffer
|
||||||
|
// which cannot fail in normal circumstances
|
||||||
|
_, _ = w.Write([]byte(n.Data))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Recursively process all child nodes
|
||||||
|
for c := n.FirstChild; c != nil; c = c.NextSibling {
|
||||||
|
extractText(w, c)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,4 +1,3 @@
|
|||||||
// Package services_test provides tests for the HTML cleaner service.
|
|
||||||
package services
|
package services
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@ -112,7 +111,7 @@ func TestHTMLCleaner_CleanHTML(t *testing.T) {
|
|||||||
{
|
{
|
||||||
name: "script and style tags content",
|
name: "script and style tags content",
|
||||||
input: "<script>alert('test');</script>Content<style>body{color:red;}</style>",
|
input: "<script>alert('test');</script>Content<style>body{color:red;}</style>",
|
||||||
expected: "alert('test');Contentbody{color:red;}",
|
expected: "Content", // Script and style tags are correctly skipped
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "line breaks and formatting",
|
name: "line breaks and formatting",
|
||||||
@ -147,7 +146,7 @@ func TestHTMLCleaner_CleanHTML(t *testing.T) {
|
|||||||
{
|
{
|
||||||
name: "special HTML5 entities",
|
name: "special HTML5 entities",
|
||||||
input: "Left arrow ← Right arrow →",
|
input: "Left arrow ← Right arrow →",
|
||||||
expected: "Left arrow ← Right arrow →", // These are not handled by the cleaner
|
expected: "Left arrow ← Right arrow →", // HTML5 entities are properly handled by the parser
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -168,7 +167,7 @@ func TestHTMLCleaner_CleanHTML_LargeContent(t *testing.T) {
|
|||||||
// Create a large HTML string
|
// Create a large HTML string
|
||||||
var builder strings.Builder
|
var builder strings.Builder
|
||||||
builder.WriteString("<html><body>")
|
builder.WriteString("<html><body>")
|
||||||
for i := 0; i < 1000; i++ {
|
for i := range 1000 {
|
||||||
builder.WriteString("<p>Paragraph ")
|
builder.WriteString("<p>Paragraph ")
|
||||||
builder.WriteString(string(rune('0' + i%10)))
|
builder.WriteString(string(rune('0' + i%10)))
|
||||||
builder.WriteString(" with some content & entities.</p>")
|
builder.WriteString(" with some content & entities.</p>")
|
||||||
@ -217,9 +216,9 @@ func TestHTMLCleaner_CleanHTML_EdgeCases(t *testing.T) {
|
|||||||
expected: "&&&",
|
expected: "&&&",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "entities without semicolon (should not be converted)",
|
name: "entities without semicolon (properly converted)",
|
||||||
input: "& test < test",
|
input: "& test < test",
|
||||||
expected: "& test < test",
|
expected: "& test < test", // Parser handles entities even without semicolons in some cases
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "mixed valid and invalid entities",
|
name: "mixed valid and invalid entities",
|
||||||
@ -234,7 +233,7 @@ func TestHTMLCleaner_CleanHTML_EdgeCases(t *testing.T) {
|
|||||||
{
|
{
|
||||||
name: "tag with no closing bracket",
|
name: "tag with no closing bracket",
|
||||||
input: "Content <p class='test' with no closing bracket",
|
input: "Content <p class='test' with no closing bracket",
|
||||||
expected: "Content <p class='test' with no closing bracket",
|
expected: "Content", // Parser handles malformed HTML gracefully
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "extremely nested tags",
|
name: "extremely nested tags",
|
||||||
@ -299,8 +298,7 @@ func BenchmarkHTMLCleaner_CleanHTML(b *testing.B) {
|
|||||||
cleaner := NewHTMLCleaner()
|
cleaner := NewHTMLCleaner()
|
||||||
input := "<div class=\"content\"><h1>Course Title</h1><p>This is a <em>great</em> course about & HTML entities like and "quotes".</p><ul><li>Item 1</li><li>Item 2</li></ul></div>"
|
input := "<div class=\"content\"><h1>Course Title</h1><p>This is a <em>great</em> course about & HTML entities like and "quotes".</p><ul><li>Item 1</li><li>Item 2</li></ul></div>"
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
cleaner.CleanHTML(input)
|
cleaner.CleanHTML(input)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -311,15 +309,14 @@ func BenchmarkHTMLCleaner_CleanHTML_Large(b *testing.B) {
|
|||||||
|
|
||||||
// Create a large HTML string
|
// Create a large HTML string
|
||||||
var builder strings.Builder
|
var builder strings.Builder
|
||||||
for i := 0; i < 100; i++ {
|
for i := range 100 {
|
||||||
builder.WriteString("<p>Paragraph ")
|
builder.WriteString("<p>Paragraph ")
|
||||||
builder.WriteString(string(rune('0' + i%10)))
|
builder.WriteString(string(rune('0' + i%10)))
|
||||||
builder.WriteString(" with some content & entities <test>.</p>")
|
builder.WriteString(" with some content & entities <test>.</p>")
|
||||||
}
|
}
|
||||||
input := builder.String()
|
input := builder.String()
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
cleaner.CleanHTML(input)
|
cleaner.CleanHTML(input)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
104
internal/services/logger.go
Normal file
104
internal/services/logger.go
Normal file
@ -0,0 +1,104 @@
|
|||||||
|
package services
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"log/slog"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/interfaces"
|
||||||
|
)
|
||||||
|
|
||||||
|
// SlogLogger implements the Logger interface using the standard library's slog package.
|
||||||
|
type SlogLogger struct {
|
||||||
|
logger *slog.Logger
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewSlogLogger creates a new structured logger using slog.
|
||||||
|
// The level parameter controls the minimum log level (debug, info, warn, error).
|
||||||
|
func NewSlogLogger(level slog.Level) interfaces.Logger {
|
||||||
|
opts := &slog.HandlerOptions{
|
||||||
|
Level: level,
|
||||||
|
}
|
||||||
|
handler := slog.NewJSONHandler(os.Stdout, opts)
|
||||||
|
return &SlogLogger{
|
||||||
|
logger: slog.New(handler),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewTextLogger creates a new structured logger with human-readable text output.
|
||||||
|
// Useful for development and debugging.
|
||||||
|
func NewTextLogger(level slog.Level) interfaces.Logger {
|
||||||
|
opts := &slog.HandlerOptions{
|
||||||
|
Level: level,
|
||||||
|
}
|
||||||
|
handler := slog.NewTextHandler(os.Stdout, opts)
|
||||||
|
return &SlogLogger{
|
||||||
|
logger: slog.New(handler),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Debug logs a debug-level message with optional key-value pairs.
|
||||||
|
func (l *SlogLogger) Debug(msg string, keysAndValues ...any) {
|
||||||
|
l.logger.Debug(msg, keysAndValues...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Info logs an info-level message with optional key-value pairs.
|
||||||
|
func (l *SlogLogger) Info(msg string, keysAndValues ...any) {
|
||||||
|
l.logger.Info(msg, keysAndValues...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Warn logs a warning-level message with optional key-value pairs.
|
||||||
|
func (l *SlogLogger) Warn(msg string, keysAndValues ...any) {
|
||||||
|
l.logger.Warn(msg, keysAndValues...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Error logs an error-level message with optional key-value pairs.
|
||||||
|
func (l *SlogLogger) Error(msg string, keysAndValues ...any) {
|
||||||
|
l.logger.Error(msg, keysAndValues...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// With returns a new logger with the given key-value pairs added as context.
|
||||||
|
func (l *SlogLogger) With(keysAndValues ...any) interfaces.Logger {
|
||||||
|
return &SlogLogger{
|
||||||
|
logger: l.logger.With(keysAndValues...),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithContext returns a new logger with context information.
|
||||||
|
// Currently preserves the logger as-is, but can be extended to extract
|
||||||
|
// trace IDs or other context values in the future.
|
||||||
|
func (l *SlogLogger) WithContext(ctx context.Context) interfaces.Logger {
|
||||||
|
// Can be extended to extract trace IDs, request IDs, etc. from context
|
||||||
|
return l
|
||||||
|
}
|
||||||
|
|
||||||
|
// NoOpLogger is a logger that discards all log messages.
|
||||||
|
// Useful for testing or when logging should be disabled.
|
||||||
|
type NoOpLogger struct{}
|
||||||
|
|
||||||
|
// NewNoOpLogger creates a logger that discards all messages.
|
||||||
|
func NewNoOpLogger() interfaces.Logger {
|
||||||
|
return &NoOpLogger{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Debug does nothing.
|
||||||
|
func (l *NoOpLogger) Debug(msg string, keysAndValues ...any) {}
|
||||||
|
|
||||||
|
// Info does nothing.
|
||||||
|
func (l *NoOpLogger) Info(msg string, keysAndValues ...any) {}
|
||||||
|
|
||||||
|
// Warn does nothing.
|
||||||
|
func (l *NoOpLogger) Warn(msg string, keysAndValues ...any) {}
|
||||||
|
|
||||||
|
// Error does nothing.
|
||||||
|
func (l *NoOpLogger) Error(msg string, keysAndValues ...any) {}
|
||||||
|
|
||||||
|
// With returns the same no-op logger.
|
||||||
|
func (l *NoOpLogger) With(keysAndValues ...any) interfaces.Logger {
|
||||||
|
return l
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithContext returns the same no-op logger.
|
||||||
|
func (l *NoOpLogger) WithContext(ctx context.Context) interfaces.Logger {
|
||||||
|
return l
|
||||||
|
}
|
||||||
95
internal/services/logger_bench_test.go
Normal file
95
internal/services/logger_bench_test.go
Normal file
@ -0,0 +1,95 @@
|
|||||||
|
package services
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"io"
|
||||||
|
"log/slog"
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
// BenchmarkSlogLogger_Info benchmarks structured JSON logging.
|
||||||
|
func BenchmarkSlogLogger_Info(b *testing.B) {
|
||||||
|
// Create logger that writes to io.Discard to avoid benchmark noise
|
||||||
|
opts := &slog.HandlerOptions{Level: slog.LevelInfo}
|
||||||
|
handler := slog.NewJSONHandler(io.Discard, opts)
|
||||||
|
logger := &SlogLogger{logger: slog.New(handler)}
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
logger.Info("test message", "key1", "value1", "key2", 42, "key3", true)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkSlogLogger_Debug benchmarks debug level logging.
|
||||||
|
func BenchmarkSlogLogger_Debug(b *testing.B) {
|
||||||
|
opts := &slog.HandlerOptions{Level: slog.LevelDebug}
|
||||||
|
handler := slog.NewJSONHandler(io.Discard, opts)
|
||||||
|
logger := &SlogLogger{logger: slog.New(handler)}
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
logger.Debug("debug message", "operation", "test", "duration", 123)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkSlogLogger_Error benchmarks error logging.
|
||||||
|
func BenchmarkSlogLogger_Error(b *testing.B) {
|
||||||
|
opts := &slog.HandlerOptions{Level: slog.LevelError}
|
||||||
|
handler := slog.NewJSONHandler(io.Discard, opts)
|
||||||
|
logger := &SlogLogger{logger: slog.New(handler)}
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
logger.Error("error occurred", "error", "test error", "code", 500)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkTextLogger_Info benchmarks text logging.
|
||||||
|
func BenchmarkTextLogger_Info(b *testing.B) {
|
||||||
|
opts := &slog.HandlerOptions{Level: slog.LevelInfo}
|
||||||
|
handler := slog.NewTextHandler(io.Discard, opts)
|
||||||
|
logger := &SlogLogger{logger: slog.New(handler)}
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
logger.Info("test message", "key1", "value1", "key2", 42)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkNoOpLogger benchmarks the no-op logger.
|
||||||
|
func BenchmarkNoOpLogger(b *testing.B) {
|
||||||
|
logger := NewNoOpLogger()
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
logger.Info("test message", "key1", "value1", "key2", 42)
|
||||||
|
logger.Error("error message", "error", "test")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkLogger_With benchmarks logger with context.
|
||||||
|
func BenchmarkLogger_With(b *testing.B) {
|
||||||
|
opts := &slog.HandlerOptions{Level: slog.LevelInfo}
|
||||||
|
handler := slog.NewJSONHandler(io.Discard, opts)
|
||||||
|
logger := &SlogLogger{logger: slog.New(handler)}
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
contextLogger := logger.With("request_id", "123", "user_id", "456")
|
||||||
|
contextLogger.Info("operation completed")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkLogger_WithContext benchmarks logger with Go context.
|
||||||
|
func BenchmarkLogger_WithContext(b *testing.B) {
|
||||||
|
opts := &slog.HandlerOptions{Level: slog.LevelInfo}
|
||||||
|
handler := slog.NewJSONHandler(io.Discard, opts)
|
||||||
|
logger := &SlogLogger{logger: slog.New(handler)}
|
||||||
|
ctx := context.Background()
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
contextLogger := logger.WithContext(ctx)
|
||||||
|
contextLogger.Info("context operation")
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -1,8 +1,7 @@
|
|||||||
// Package services provides the core functionality for the articulate-parser application.
|
|
||||||
// It implements the interfaces defined in the interfaces package.
|
|
||||||
package services
|
package services
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
@ -16,6 +15,9 @@ import (
|
|||||||
"github.com/kjanat/articulate-parser/internal/models"
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// shareIDRegex is compiled once at package init for extracting share IDs from URIs.
|
||||||
|
var shareIDRegex = regexp.MustCompile(`/share/([a-zA-Z0-9_-]+)`)
|
||||||
|
|
||||||
// ArticulateParser implements the CourseParser interface specifically for Articulate Rise courses.
|
// ArticulateParser implements the CourseParser interface specifically for Articulate Rise courses.
|
||||||
// It can fetch courses from the Articulate Rise API or load them from local JSON files.
|
// It can fetch courses from the Articulate Rise API or load them from local JSON files.
|
||||||
type ArticulateParser struct {
|
type ArticulateParser struct {
|
||||||
@ -23,32 +25,36 @@ type ArticulateParser struct {
|
|||||||
BaseURL string
|
BaseURL string
|
||||||
// Client is the HTTP client used to make requests to the API
|
// Client is the HTTP client used to make requests to the API
|
||||||
Client *http.Client
|
Client *http.Client
|
||||||
|
// Logger for structured logging
|
||||||
|
Logger interfaces.Logger
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewArticulateParser creates a new ArticulateParser instance with default settings.
|
// NewArticulateParser creates a new ArticulateParser instance.
|
||||||
// The default configuration uses the standard Articulate Rise API URL and a
|
// If baseURL is empty, uses the default Articulate Rise API URL.
|
||||||
// HTTP client with a 30-second timeout.
|
// If timeout is zero, uses a 30-second timeout.
|
||||||
func NewArticulateParser() interfaces.CourseParser {
|
func NewArticulateParser(logger interfaces.Logger, baseURL string, timeout time.Duration) interfaces.CourseParser {
|
||||||
|
if logger == nil {
|
||||||
|
logger = NewNoOpLogger()
|
||||||
|
}
|
||||||
|
if baseURL == "" {
|
||||||
|
baseURL = "https://rise.articulate.com"
|
||||||
|
}
|
||||||
|
if timeout == 0 {
|
||||||
|
timeout = 30 * time.Second
|
||||||
|
}
|
||||||
return &ArticulateParser{
|
return &ArticulateParser{
|
||||||
BaseURL: "https://rise.articulate.com",
|
BaseURL: baseURL,
|
||||||
Client: &http.Client{
|
Client: &http.Client{
|
||||||
Timeout: 30 * time.Second,
|
Timeout: timeout,
|
||||||
},
|
},
|
||||||
|
Logger: logger,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// FetchCourse fetches a course from the given URI.
|
// FetchCourse fetches a course from the given URI and returns the parsed course data.
|
||||||
// It extracts the share ID from the URI, constructs an API URL, and fetches the course data.
|
// The URI should be an Articulate Rise share URL (e.g., https://rise.articulate.com/share/SHARE_ID).
|
||||||
// The course data is then unmarshalled into a Course model.
|
// The context can be used for cancellation and timeout control.
|
||||||
//
|
func (p *ArticulateParser) FetchCourse(ctx context.Context, uri string) (*models.Course, error) {
|
||||||
// Parameters:
|
|
||||||
// - uri: The Articulate Rise share URL (e.g., https://rise.articulate.com/share/SHARE_ID)
|
|
||||||
//
|
|
||||||
// Returns:
|
|
||||||
// - A parsed Course model if successful
|
|
||||||
// - An error if the fetch fails, if the share ID can't be extracted,
|
|
||||||
// or if the response can't be parsed
|
|
||||||
func (p *ArticulateParser) FetchCourse(uri string) (*models.Course, error) {
|
|
||||||
shareID, err := p.extractShareID(uri)
|
shareID, err := p.extractShareID(uri)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@ -56,21 +62,34 @@ func (p *ArticulateParser) FetchCourse(uri string) (*models.Course, error) {
|
|||||||
|
|
||||||
apiURL := p.buildAPIURL(shareID)
|
apiURL := p.buildAPIURL(shareID)
|
||||||
|
|
||||||
resp, err := p.Client.Get(apiURL)
|
req, err := http.NewRequestWithContext(ctx, http.MethodGet, apiURL, http.NoBody)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to create request: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := p.Client.Do(req)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("failed to fetch course data: %w", err)
|
return nil, fmt.Errorf("failed to fetch course data: %w", err)
|
||||||
}
|
}
|
||||||
defer resp.Body.Close()
|
// Ensure response body is closed even if ReadAll fails. Close errors are logged
|
||||||
|
// but not fatal since the body content has already been read and parsed. In the
|
||||||
if resp.StatusCode != http.StatusOK {
|
// context of HTTP responses, the body must be closed to release the underlying
|
||||||
return nil, fmt.Errorf("API returned status %d", resp.StatusCode)
|
// connection, but a close error doesn't invalidate the data already consumed.
|
||||||
}
|
defer func() {
|
||||||
|
if err := resp.Body.Close(); err != nil {
|
||||||
|
p.Logger.Warn("failed to close response body", "error", err, "url", apiURL)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
body, err := io.ReadAll(resp.Body)
|
body, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("failed to read response body: %w", err)
|
return nil, fmt.Errorf("failed to read response body: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if resp.StatusCode != http.StatusOK {
|
||||||
|
return nil, fmt.Errorf("API returned status %d: %s", resp.StatusCode, string(body))
|
||||||
|
}
|
||||||
|
|
||||||
var course models.Course
|
var course models.Course
|
||||||
if err := json.Unmarshal(body, &course); err != nil {
|
if err := json.Unmarshal(body, &course); err != nil {
|
||||||
return nil, fmt.Errorf("failed to unmarshal JSON: %w", err)
|
return nil, fmt.Errorf("failed to unmarshal JSON: %w", err)
|
||||||
@ -80,15 +99,8 @@ func (p *ArticulateParser) FetchCourse(uri string) (*models.Course, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// LoadCourseFromFile loads an Articulate Rise course from a local JSON file.
|
// LoadCourseFromFile loads an Articulate Rise course from a local JSON file.
|
||||||
// The file should contain a valid JSON representation of an Articulate Rise course.
|
|
||||||
//
|
|
||||||
// Parameters:
|
|
||||||
// - filePath: The path to the JSON file containing the course data
|
|
||||||
//
|
|
||||||
// Returns:
|
|
||||||
// - A parsed Course model if successful
|
|
||||||
// - An error if the file can't be read or the JSON can't be parsed
|
|
||||||
func (p *ArticulateParser) LoadCourseFromFile(filePath string) (*models.Course, error) {
|
func (p *ArticulateParser) LoadCourseFromFile(filePath string) (*models.Course, error) {
|
||||||
|
// #nosec G304 - File path is provided by user via CLI argument, which is expected behavior
|
||||||
data, err := os.ReadFile(filePath)
|
data, err := os.ReadFile(filePath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("failed to read file: %w", err)
|
return nil, fmt.Errorf("failed to read file: %w", err)
|
||||||
@ -124,8 +136,7 @@ func (p *ArticulateParser) extractShareID(uri string) (string, error) {
|
|||||||
return "", fmt.Errorf("invalid domain for Articulate Rise URI: %s", parsedURL.Host)
|
return "", fmt.Errorf("invalid domain for Articulate Rise URI: %s", parsedURL.Host)
|
||||||
}
|
}
|
||||||
|
|
||||||
re := regexp.MustCompile(`/share/([a-zA-Z0-9_-]+)`)
|
matches := shareIDRegex.FindStringSubmatch(uri)
|
||||||
matches := re.FindStringSubmatch(uri)
|
|
||||||
if len(matches) < 2 {
|
if len(matches) < 2 {
|
||||||
return "", fmt.Errorf("could not extract share ID from URI: %s", uri)
|
return "", fmt.Errorf("could not extract share ID from URI: %s", uri)
|
||||||
}
|
}
|
||||||
|
|||||||
219
internal/services/parser_bench_test.go
Normal file
219
internal/services/parser_bench_test.go
Normal file
@ -0,0 +1,219 @@
|
|||||||
|
package services
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"encoding/json"
|
||||||
|
"net/http"
|
||||||
|
"net/http/httptest"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
|
)
|
||||||
|
|
||||||
|
// BenchmarkArticulateParser_FetchCourse benchmarks the FetchCourse method.
|
||||||
|
func BenchmarkArticulateParser_FetchCourse(b *testing.B) {
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "benchmark-id",
|
||||||
|
Author: "Benchmark Author",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
ID: "bench-course",
|
||||||
|
Title: "Benchmark Course",
|
||||||
|
Description: "Testing performance",
|
||||||
|
Lessons: []models.Lesson{
|
||||||
|
{
|
||||||
|
ID: "lesson1",
|
||||||
|
Title: "Lesson 1",
|
||||||
|
Type: "lesson",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
w.Header().Set("Content-Type", "application/json")
|
||||||
|
// Encode errors are ignored in benchmarks; the test server's ResponseWriter
|
||||||
|
// writes are reliable and any encoding error would be a test setup issue
|
||||||
|
_ = json.NewEncoder(w).Encode(testCourse)
|
||||||
|
}))
|
||||||
|
defer server.Close()
|
||||||
|
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: server.URL,
|
||||||
|
Client: &http.Client{},
|
||||||
|
Logger: NewNoOpLogger(),
|
||||||
|
}
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
_, err := parser.FetchCourse(context.Background(), "https://rise.articulate.com/share/benchmark-id")
|
||||||
|
if err != nil {
|
||||||
|
b.Fatalf("FetchCourse failed: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkArticulateParser_FetchCourse_LargeCourse benchmarks with a large course.
|
||||||
|
func BenchmarkArticulateParser_FetchCourse_LargeCourse(b *testing.B) {
|
||||||
|
// Create a large course with many lessons
|
||||||
|
lessons := make([]models.Lesson, 100)
|
||||||
|
for i := range 100 {
|
||||||
|
lessons[i] = models.Lesson{
|
||||||
|
ID: string(rune(i)),
|
||||||
|
Title: "Lesson " + string(rune(i)),
|
||||||
|
Type: "lesson",
|
||||||
|
Description: "This is a test lesson with some description",
|
||||||
|
Items: []models.Item{
|
||||||
|
{
|
||||||
|
Type: "text",
|
||||||
|
Items: []models.SubItem{
|
||||||
|
{
|
||||||
|
Heading: "Test Heading",
|
||||||
|
Paragraph: "Test paragraph content with some text",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "large-course-id",
|
||||||
|
Author: "Benchmark Author",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
ID: "large-course",
|
||||||
|
Title: "Large Benchmark Course",
|
||||||
|
Description: "Testing performance with large course",
|
||||||
|
Lessons: lessons,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
w.Header().Set("Content-Type", "application/json")
|
||||||
|
// Encode errors are ignored in benchmarks; the test server's ResponseWriter
|
||||||
|
// writes are reliable and any encoding error would be a test setup issue
|
||||||
|
_ = json.NewEncoder(w).Encode(testCourse)
|
||||||
|
}))
|
||||||
|
defer server.Close()
|
||||||
|
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: server.URL,
|
||||||
|
Client: &http.Client{},
|
||||||
|
Logger: NewNoOpLogger(),
|
||||||
|
}
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
_, err := parser.FetchCourse(context.Background(), "https://rise.articulate.com/share/large-course-id")
|
||||||
|
if err != nil {
|
||||||
|
b.Fatalf("FetchCourse failed: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkArticulateParser_LoadCourseFromFile benchmarks loading from file.
|
||||||
|
func BenchmarkArticulateParser_LoadCourseFromFile(b *testing.B) {
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "file-test-id",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
Title: "File Test Course",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
tempDir := b.TempDir()
|
||||||
|
tempFile := filepath.Join(tempDir, "benchmark.json")
|
||||||
|
|
||||||
|
data, err := json.Marshal(testCourse)
|
||||||
|
if err != nil {
|
||||||
|
b.Fatalf("Failed to marshal: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := os.WriteFile(tempFile, data, 0o644); err != nil {
|
||||||
|
b.Fatalf("Failed to write file: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
parser := NewArticulateParser(nil, "", 0)
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
_, err := parser.LoadCourseFromFile(tempFile)
|
||||||
|
if err != nil {
|
||||||
|
b.Fatalf("LoadCourseFromFile failed: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkArticulateParser_LoadCourseFromFile_Large benchmarks with large file.
|
||||||
|
func BenchmarkArticulateParser_LoadCourseFromFile_Large(b *testing.B) {
|
||||||
|
// Create a large course
|
||||||
|
lessons := make([]models.Lesson, 200)
|
||||||
|
for i := range 200 {
|
||||||
|
lessons[i] = models.Lesson{
|
||||||
|
ID: string(rune(i)),
|
||||||
|
Title: "Lesson " + string(rune(i)),
|
||||||
|
Type: "lesson",
|
||||||
|
Items: []models.Item{
|
||||||
|
{Type: "text", Items: []models.SubItem{{Heading: "H", Paragraph: "P"}}},
|
||||||
|
{Type: "list", Items: []models.SubItem{{Paragraph: "Item 1"}, {Paragraph: "Item 2"}}},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "large-file-id",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
Title: "Large File Course",
|
||||||
|
Lessons: lessons,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
tempDir := b.TempDir()
|
||||||
|
tempFile := filepath.Join(tempDir, "large-benchmark.json")
|
||||||
|
|
||||||
|
data, err := json.Marshal(testCourse)
|
||||||
|
if err != nil {
|
||||||
|
b.Fatalf("Failed to marshal: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := os.WriteFile(tempFile, data, 0o644); err != nil {
|
||||||
|
b.Fatalf("Failed to write file: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
parser := NewArticulateParser(nil, "", 0)
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
_, err := parser.LoadCourseFromFile(tempFile)
|
||||||
|
if err != nil {
|
||||||
|
b.Fatalf("LoadCourseFromFile failed: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkArticulateParser_ExtractShareID benchmarks share ID extraction.
|
||||||
|
func BenchmarkArticulateParser_ExtractShareID(b *testing.B) {
|
||||||
|
parser := &ArticulateParser{}
|
||||||
|
uri := "https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/"
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
_, err := parser.extractShareID(uri)
|
||||||
|
if err != nil {
|
||||||
|
b.Fatalf("extractShareID failed: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BenchmarkArticulateParser_BuildAPIURL benchmarks API URL building.
|
||||||
|
func BenchmarkArticulateParser_BuildAPIURL(b *testing.B) {
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: "https://rise.articulate.com",
|
||||||
|
}
|
||||||
|
shareID := "test-share-id-12345"
|
||||||
|
|
||||||
|
b.ResetTimer()
|
||||||
|
for b.Loop() {
|
||||||
|
_ = parser.buildAPIURL(shareID)
|
||||||
|
}
|
||||||
|
}
|
||||||
289
internal/services/parser_context_test.go
Normal file
289
internal/services/parser_context_test.go
Normal file
@ -0,0 +1,289 @@
|
|||||||
|
package services
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"encoding/json"
|
||||||
|
"net/http"
|
||||||
|
"net/http/httptest"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/models"
|
||||||
|
)
|
||||||
|
|
||||||
|
// TestArticulateParser_FetchCourse_ContextCancellation tests that FetchCourse
|
||||||
|
// respects context cancellation.
|
||||||
|
func TestArticulateParser_FetchCourse_ContextCancellation(t *testing.T) {
|
||||||
|
// Create a server that delays response
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
// Sleep to give time for context cancellation
|
||||||
|
time.Sleep(100 * time.Millisecond)
|
||||||
|
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "test-id",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
Title: "Test Course",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
// Encode errors are ignored in test setup; httptest.ResponseWriter is reliable
|
||||||
|
_ = json.NewEncoder(w).Encode(testCourse)
|
||||||
|
}))
|
||||||
|
defer server.Close()
|
||||||
|
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: server.URL,
|
||||||
|
Client: &http.Client{
|
||||||
|
Timeout: 5 * time.Second,
|
||||||
|
},
|
||||||
|
Logger: NewNoOpLogger(),
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create a context that we'll cancel immediately
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
cancel() // Cancel immediately
|
||||||
|
|
||||||
|
_, err := parser.FetchCourse(ctx, "https://rise.articulate.com/share/test-id")
|
||||||
|
|
||||||
|
// Should get a context cancellation error
|
||||||
|
if err == nil {
|
||||||
|
t.Fatal("Expected error due to context cancellation, got nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
if !strings.Contains(err.Error(), "context canceled") {
|
||||||
|
t.Errorf("Expected context cancellation error, got: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestArticulateParser_FetchCourse_ContextTimeout tests that FetchCourse
|
||||||
|
// respects context timeout.
|
||||||
|
func TestArticulateParser_FetchCourse_ContextTimeout(t *testing.T) {
|
||||||
|
// Create a server that delays response longer than timeout
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
// Sleep longer than the context timeout
|
||||||
|
time.Sleep(200 * time.Millisecond)
|
||||||
|
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "test-id",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
Title: "Test Course",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
// Encode errors are ignored in test setup; httptest.ResponseWriter is reliable
|
||||||
|
_ = json.NewEncoder(w).Encode(testCourse)
|
||||||
|
}))
|
||||||
|
defer server.Close()
|
||||||
|
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: server.URL,
|
||||||
|
Client: &http.Client{
|
||||||
|
Timeout: 5 * time.Second,
|
||||||
|
},
|
||||||
|
Logger: NewNoOpLogger(),
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create a context with a very short timeout
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), 50*time.Millisecond)
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
_, err := parser.FetchCourse(ctx, "https://rise.articulate.com/share/test-id")
|
||||||
|
|
||||||
|
// Should get a context deadline exceeded error
|
||||||
|
if err == nil {
|
||||||
|
t.Fatal("Expected error due to context timeout, got nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
if !strings.Contains(err.Error(), "deadline exceeded") &&
|
||||||
|
!strings.Contains(err.Error(), "context deadline exceeded") {
|
||||||
|
t.Errorf("Expected context timeout error, got: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestArticulateParser_FetchCourse_ContextDeadline tests that FetchCourse
|
||||||
|
// respects context deadline.
|
||||||
|
func TestArticulateParser_FetchCourse_ContextDeadline(t *testing.T) {
|
||||||
|
// Create a server that delays response
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
time.Sleep(150 * time.Millisecond)
|
||||||
|
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "test-id",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
Title: "Test Course",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
// Encode errors are ignored in test setup; httptest.ResponseWriter is reliable
|
||||||
|
_ = json.NewEncoder(w).Encode(testCourse)
|
||||||
|
}))
|
||||||
|
defer server.Close()
|
||||||
|
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: server.URL,
|
||||||
|
Client: &http.Client{
|
||||||
|
Timeout: 5 * time.Second,
|
||||||
|
},
|
||||||
|
Logger: NewNoOpLogger(),
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create a context with a deadline in the past
|
||||||
|
ctx, cancel := context.WithDeadline(context.Background(), time.Now().Add(10*time.Millisecond))
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
_, err := parser.FetchCourse(ctx, "https://rise.articulate.com/share/test-id")
|
||||||
|
|
||||||
|
// Should get a deadline exceeded error
|
||||||
|
if err == nil {
|
||||||
|
t.Fatal("Expected error due to context deadline, got nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
if !strings.Contains(err.Error(), "deadline exceeded") &&
|
||||||
|
!strings.Contains(err.Error(), "context deadline exceeded") {
|
||||||
|
t.Errorf("Expected deadline exceeded error, got: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestArticulateParser_FetchCourse_ContextSuccess tests that FetchCourse
|
||||||
|
// succeeds when context is not canceled.
|
||||||
|
func TestArticulateParser_FetchCourse_ContextSuccess(t *testing.T) {
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "test-id",
|
||||||
|
Course: models.CourseInfo{
|
||||||
|
Title: "Test Course",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
// Respond quickly
|
||||||
|
// Encode errors are ignored in test setup; httptest.ResponseWriter is reliable
|
||||||
|
_ = json.NewEncoder(w).Encode(testCourse)
|
||||||
|
}))
|
||||||
|
defer server.Close()
|
||||||
|
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: server.URL,
|
||||||
|
Client: &http.Client{
|
||||||
|
Timeout: 5 * time.Second,
|
||||||
|
},
|
||||||
|
Logger: NewNoOpLogger(),
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create a context with generous timeout
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second)
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
course, err := parser.FetchCourse(ctx, "https://rise.articulate.com/share/test-id")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("Expected no error, got: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if course == nil {
|
||||||
|
t.Fatal("Expected course, got nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
if course.Course.Title != testCourse.Course.Title {
|
||||||
|
t.Errorf("Expected title '%s', got '%s'", testCourse.Course.Title, course.Course.Title)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestArticulateParser_FetchCourse_CancellationDuringRequest tests cancellation
|
||||||
|
// during an in-flight request.
|
||||||
|
func TestArticulateParser_FetchCourse_CancellationDuringRequest(t *testing.T) {
|
||||||
|
requestStarted := make(chan bool)
|
||||||
|
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
requestStarted <- true
|
||||||
|
// Keep the handler running to simulate slow response
|
||||||
|
time.Sleep(300 * time.Millisecond)
|
||||||
|
|
||||||
|
testCourse := &models.Course{
|
||||||
|
ShareID: "test-id",
|
||||||
|
}
|
||||||
|
// Encode errors are ignored in test setup; httptest.ResponseWriter is reliable
|
||||||
|
_ = json.NewEncoder(w).Encode(testCourse)
|
||||||
|
}))
|
||||||
|
defer server.Close()
|
||||||
|
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: server.URL,
|
||||||
|
Client: &http.Client{
|
||||||
|
Timeout: 5 * time.Second,
|
||||||
|
},
|
||||||
|
Logger: NewNoOpLogger(),
|
||||||
|
}
|
||||||
|
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
|
||||||
|
// Start the request in a goroutine
|
||||||
|
errChan := make(chan error, 1)
|
||||||
|
go func() {
|
||||||
|
_, err := parser.FetchCourse(ctx, "https://rise.articulate.com/share/test-id")
|
||||||
|
errChan <- err
|
||||||
|
}()
|
||||||
|
|
||||||
|
// Wait for request to start
|
||||||
|
<-requestStarted
|
||||||
|
|
||||||
|
// Cancel after request has started
|
||||||
|
cancel()
|
||||||
|
|
||||||
|
// Get the error
|
||||||
|
err := <-errChan
|
||||||
|
|
||||||
|
if err == nil {
|
||||||
|
t.Fatal("Expected error due to context cancellation, got nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Should contain context canceled somewhere in the error chain
|
||||||
|
if !strings.Contains(err.Error(), "context canceled") {
|
||||||
|
t.Errorf("Expected context canceled error, got: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestArticulateParser_FetchCourse_MultipleTimeouts tests behavior with
|
||||||
|
// multiple concurrent requests and timeouts.
|
||||||
|
func TestArticulateParser_FetchCourse_MultipleTimeouts(t *testing.T) {
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
time.Sleep(100 * time.Millisecond)
|
||||||
|
testCourse := &models.Course{ShareID: "test"}
|
||||||
|
// Encode errors are ignored in test setup; httptest.ResponseWriter is reliable
|
||||||
|
_ = json.NewEncoder(w).Encode(testCourse)
|
||||||
|
}))
|
||||||
|
defer server.Close()
|
||||||
|
|
||||||
|
parser := &ArticulateParser{
|
||||||
|
BaseURL: server.URL,
|
||||||
|
Client: &http.Client{
|
||||||
|
Timeout: 5 * time.Second,
|
||||||
|
},
|
||||||
|
Logger: NewNoOpLogger(),
|
||||||
|
}
|
||||||
|
|
||||||
|
// Launch multiple requests with different timeouts
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
timeout time.Duration
|
||||||
|
shouldSucceed bool
|
||||||
|
}{
|
||||||
|
{"very short timeout", 10 * time.Millisecond, false},
|
||||||
|
{"short timeout", 50 * time.Millisecond, false},
|
||||||
|
{"adequate timeout", 500 * time.Millisecond, true},
|
||||||
|
{"long timeout", 2 * time.Second, true},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), tt.timeout)
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
_, err := parser.FetchCourse(ctx, "https://rise.articulate.com/share/test-id")
|
||||||
|
|
||||||
|
if tt.shouldSucceed && err != nil {
|
||||||
|
t.Errorf("Expected success with timeout %v, got error: %v", tt.timeout, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if !tt.shouldSucceed && err == nil {
|
||||||
|
t.Errorf("Expected timeout error with timeout %v, got success", tt.timeout)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -1,7 +1,7 @@
|
|||||||
// Package services_test provides tests for the parser service.
|
|
||||||
package services
|
package services
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/http/httptest"
|
"net/http/httptest"
|
||||||
@ -16,7 +16,7 @@ import (
|
|||||||
|
|
||||||
// TestNewArticulateParser tests the NewArticulateParser constructor.
|
// TestNewArticulateParser tests the NewArticulateParser constructor.
|
||||||
func TestNewArticulateParser(t *testing.T) {
|
func TestNewArticulateParser(t *testing.T) {
|
||||||
parser := NewArticulateParser()
|
parser := NewArticulateParser(nil, "", 0)
|
||||||
|
|
||||||
if parser == nil {
|
if parser == nil {
|
||||||
t.Fatal("NewArticulateParser() returned nil")
|
t.Fatal("NewArticulateParser() returned nil")
|
||||||
@ -112,7 +112,7 @@ func TestArticulateParser_FetchCourse(t *testing.T) {
|
|||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
course, err := parser.FetchCourse(tt.uri)
|
course, err := parser.FetchCourse(context.Background(), tt.uri)
|
||||||
|
|
||||||
if tt.expectedError != "" {
|
if tt.expectedError != "" {
|
||||||
if err == nil {
|
if err == nil {
|
||||||
@ -146,7 +146,7 @@ func TestArticulateParser_FetchCourse_NetworkError(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
_, err := parser.FetchCourse("https://rise.articulate.com/share/test-share-id")
|
_, err := parser.FetchCourse(context.Background(), "https://rise.articulate.com/share/test-share-id")
|
||||||
if err == nil {
|
if err == nil {
|
||||||
t.Fatal("Expected network error, got nil")
|
t.Fatal("Expected network error, got nil")
|
||||||
}
|
}
|
||||||
@ -161,7 +161,10 @@ func TestArticulateParser_FetchCourse_InvalidJSON(t *testing.T) {
|
|||||||
// Create test server that returns invalid JSON
|
// Create test server that returns invalid JSON
|
||||||
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
w.Header().Set("Content-Type", "application/json")
|
w.Header().Set("Content-Type", "application/json")
|
||||||
w.Write([]byte("invalid json"))
|
// Write is used for its side effect; the test verifies error handling on
|
||||||
|
// the client side, not whether the write succeeds. Ignore the error since
|
||||||
|
// httptest.ResponseWriter writes are rarely problematic in test contexts.
|
||||||
|
_, _ = w.Write([]byte("invalid json"))
|
||||||
}))
|
}))
|
||||||
defer server.Close()
|
defer server.Close()
|
||||||
|
|
||||||
@ -172,7 +175,7 @@ func TestArticulateParser_FetchCourse_InvalidJSON(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
_, err := parser.FetchCourse("https://rise.articulate.com/share/test-share-id")
|
_, err := parser.FetchCourse(context.Background(), "https://rise.articulate.com/share/test-share-id")
|
||||||
if err == nil {
|
if err == nil {
|
||||||
t.Fatal("Expected JSON parsing error, got nil")
|
t.Fatal("Expected JSON parsing error, got nil")
|
||||||
}
|
}
|
||||||
@ -205,11 +208,11 @@ func TestArticulateParser_LoadCourseFromFile(t *testing.T) {
|
|||||||
t.Fatalf("Failed to marshal test course: %v", err)
|
t.Fatalf("Failed to marshal test course: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := os.WriteFile(tempFile, data, 0644); err != nil {
|
if err := os.WriteFile(tempFile, data, 0o644); err != nil {
|
||||||
t.Fatalf("Failed to write test file: %v", err)
|
t.Fatalf("Failed to write test file: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
parser := NewArticulateParser()
|
parser := NewArticulateParser(nil, "", 0)
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
@ -264,11 +267,11 @@ func TestArticulateParser_LoadCourseFromFile_InvalidJSON(t *testing.T) {
|
|||||||
tempDir := t.TempDir()
|
tempDir := t.TempDir()
|
||||||
tempFile := filepath.Join(tempDir, "invalid.json")
|
tempFile := filepath.Join(tempDir, "invalid.json")
|
||||||
|
|
||||||
if err := os.WriteFile(tempFile, []byte("invalid json content"), 0644); err != nil {
|
if err := os.WriteFile(tempFile, []byte("invalid json content"), 0o644); err != nil {
|
||||||
t.Fatalf("Failed to write test file: %v", err)
|
t.Fatalf("Failed to write test file: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
parser := NewArticulateParser()
|
parser := NewArticulateParser(nil, "", 0)
|
||||||
_, err := parser.LoadCourseFromFile(tempFile)
|
_, err := parser.LoadCourseFromFile(tempFile)
|
||||||
|
|
||||||
if err == nil {
|
if err == nil {
|
||||||
@ -420,8 +423,7 @@ func BenchmarkExtractShareID(b *testing.B) {
|
|||||||
parser := &ArticulateParser{}
|
parser := &ArticulateParser{}
|
||||||
uri := "https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/"
|
uri := "https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/"
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
_, _ = parser.extractShareID(uri)
|
_, _ = parser.extractShareID(uri)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -433,8 +435,7 @@ func BenchmarkBuildAPIURL(b *testing.B) {
|
|||||||
}
|
}
|
||||||
shareID := "N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO"
|
shareID := "N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO"
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
_ = parser.buildAPIURL(shareID)
|
_ = parser.buildAPIURL(shareID)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -5,7 +5,17 @@ package version
|
|||||||
// Version information.
|
// Version information.
|
||||||
var (
|
var (
|
||||||
// Version is the current version of the application.
|
// Version is the current version of the application.
|
||||||
Version = "0.4.1"
|
// Breaking changes from 0.4.x:
|
||||||
|
// - Renamed GetSupportedFormat() -> SupportedFormat()
|
||||||
|
// - Renamed GetSupportedFormats() -> SupportedFormats()
|
||||||
|
// - FetchCourse now requires context.Context parameter
|
||||||
|
// - NewArticulateParser now accepts logger, baseURL, timeout
|
||||||
|
// New features:
|
||||||
|
// - Structured logging with slog
|
||||||
|
// - Configuration via environment variables
|
||||||
|
// - Context-aware HTTP requests
|
||||||
|
// - Comprehensive benchmarks and examples.
|
||||||
|
Version = "1.0.0"
|
||||||
|
|
||||||
// BuildTime is the time the binary was built.
|
// BuildTime is the time the binary was built.
|
||||||
BuildTime = "unknown"
|
BuildTime = "unknown"
|
||||||
|
|||||||
30
main.go
30
main.go
@ -4,12 +4,14 @@
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"log"
|
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"github.com/kjanat/articulate-parser/internal/config"
|
||||||
"github.com/kjanat/articulate-parser/internal/exporters"
|
"github.com/kjanat/articulate-parser/internal/exporters"
|
||||||
|
"github.com/kjanat/articulate-parser/internal/interfaces"
|
||||||
"github.com/kjanat/articulate-parser/internal/services"
|
"github.com/kjanat/articulate-parser/internal/services"
|
||||||
"github.com/kjanat/articulate-parser/internal/version"
|
"github.com/kjanat/articulate-parser/internal/version"
|
||||||
)
|
)
|
||||||
@ -24,9 +26,19 @@ func main() {
|
|||||||
// run contains the main application logic and returns an exit code.
|
// run contains the main application logic and returns an exit code.
|
||||||
// This function is testable as it doesn't call os.Exit directly.
|
// This function is testable as it doesn't call os.Exit directly.
|
||||||
func run(args []string) int {
|
func run(args []string) int {
|
||||||
// Dependency injection setup
|
// Load configuration
|
||||||
|
cfg := config.Load()
|
||||||
|
|
||||||
|
// Dependency injection setup with configuration
|
||||||
|
var logger interfaces.Logger
|
||||||
|
if cfg.LogFormat == "json" {
|
||||||
|
logger = services.NewSlogLogger(cfg.LogLevel)
|
||||||
|
} else {
|
||||||
|
logger = services.NewTextLogger(cfg.LogLevel)
|
||||||
|
}
|
||||||
|
|
||||||
htmlCleaner := services.NewHTMLCleaner()
|
htmlCleaner := services.NewHTMLCleaner()
|
||||||
parser := services.NewArticulateParser()
|
parser := services.NewArticulateParser(logger, cfg.BaseURL, cfg.RequestTimeout)
|
||||||
exporterFactory := exporters.NewFactory(htmlCleaner)
|
exporterFactory := exporters.NewFactory(htmlCleaner)
|
||||||
app := services.NewApp(parser, exporterFactory)
|
app := services.NewApp(parser, exporterFactory)
|
||||||
|
|
||||||
@ -40,13 +52,13 @@ func run(args []string) int {
|
|||||||
|
|
||||||
// Check for help flag
|
// Check for help flag
|
||||||
if len(args) > 1 && (args[1] == "--help" || args[1] == "-h" || args[1] == "help") {
|
if len(args) > 1 && (args[1] == "--help" || args[1] == "-h" || args[1] == "help") {
|
||||||
printUsage(args[0], app.GetSupportedFormats())
|
printUsage(args[0], app.SupportedFormats())
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for required command-line arguments
|
// Check for required command-line arguments
|
||||||
if len(args) < 4 {
|
if len(args) < 4 {
|
||||||
printUsage(args[0], app.GetSupportedFormats())
|
printUsage(args[0], app.SupportedFormats())
|
||||||
return 1
|
return 1
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -58,17 +70,17 @@ func run(args []string) int {
|
|||||||
|
|
||||||
// Determine if source is a URI or file path
|
// Determine if source is a URI or file path
|
||||||
if isURI(source) {
|
if isURI(source) {
|
||||||
err = app.ProcessCourseFromURI(source, format, output)
|
err = app.ProcessCourseFromURI(context.Background(), source, format, output)
|
||||||
} else {
|
} else {
|
||||||
err = app.ProcessCourseFromFile(source, format, output)
|
err = app.ProcessCourseFromFile(source, format, output)
|
||||||
}
|
}
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Printf("Error processing course: %v", err)
|
logger.Error("failed to process course", "error", err, "source", source)
|
||||||
return 1
|
return 1
|
||||||
}
|
}
|
||||||
|
|
||||||
fmt.Printf("Successfully exported course to %s\n", output)
|
logger.Info("successfully exported course", "output", output, "format", format)
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -80,7 +92,7 @@ func run(args []string) int {
|
|||||||
// Returns:
|
// Returns:
|
||||||
// - true if the string appears to be a URI, false otherwise
|
// - true if the string appears to be a URI, false otherwise
|
||||||
func isURI(str string) bool {
|
func isURI(str string) bool {
|
||||||
return len(str) > 7 && (str[:7] == "http://" || str[:8] == "https://")
|
return strings.HasPrefix(str, "http://") || strings.HasPrefix(str, "https://")
|
||||||
}
|
}
|
||||||
|
|
||||||
// printUsage prints the command-line usage information.
|
// printUsage prints the command-line usage information.
|
||||||
|
|||||||
153
main_test.go
153
main_test.go
@ -1,9 +1,7 @@
|
|||||||
// Package main_test provides tests for the main package utility functions.
|
|
||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"fmt"
|
|
||||||
"io"
|
"io"
|
||||||
"log"
|
"log"
|
||||||
"os"
|
"os"
|
||||||
@ -89,8 +87,7 @@ func TestIsURI(t *testing.T) {
|
|||||||
func BenchmarkIsURI(b *testing.B) {
|
func BenchmarkIsURI(b *testing.B) {
|
||||||
testStr := "https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/"
|
testStr := "https://rise.articulate.com/share/N_APNg40Vr2CSH2xNz-ZLATM5kNviDIO#/"
|
||||||
|
|
||||||
b.ResetTimer()
|
for b.Loop() {
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
isURI(testStr)
|
isURI(testStr)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -125,13 +122,16 @@ func TestRunWithInsufficientArgs(t *testing.T) {
|
|||||||
// Run the function
|
// Run the function
|
||||||
exitCode := run(tt.args)
|
exitCode := run(tt.args)
|
||||||
|
|
||||||
// Restore stdout
|
// Restore stdout. Close errors are ignored: we've already captured the
|
||||||
w.Close()
|
// output before closing, and any close error doesn't affect test validity.
|
||||||
|
_ = w.Close()
|
||||||
os.Stdout = oldStdout
|
os.Stdout = oldStdout
|
||||||
|
|
||||||
// Read captured output
|
// Read captured output. Copy errors are ignored: in this test context,
|
||||||
|
// reading from a pipe that was just closed is not expected to fail, and
|
||||||
|
// we're verifying the captured output regardless.
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
io.Copy(&buf, r)
|
_, _ = io.Copy(&buf, r)
|
||||||
output := buf.String()
|
output := buf.String()
|
||||||
|
|
||||||
// Verify exit code
|
// Verify exit code
|
||||||
@ -166,13 +166,15 @@ func TestRunWithHelpFlags(t *testing.T) {
|
|||||||
args := []string{"articulate-parser", flag}
|
args := []string{"articulate-parser", flag}
|
||||||
exitCode := run(args)
|
exitCode := run(args)
|
||||||
|
|
||||||
// Restore stdout
|
// Restore stdout. Close errors are ignored: the pipe write end is already
|
||||||
w.Close()
|
// closed before reading, and any close error doesn't affect the test.
|
||||||
|
_ = w.Close()
|
||||||
os.Stdout = oldStdout
|
os.Stdout = oldStdout
|
||||||
|
|
||||||
// Read captured output
|
// Read captured output. Copy errors are ignored: we successfully wrote
|
||||||
|
// the help output to the pipe and can verify it regardless of close semantics.
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
io.Copy(&buf, r)
|
_, _ = io.Copy(&buf, r)
|
||||||
output := buf.String()
|
output := buf.String()
|
||||||
|
|
||||||
// Verify exit code is 0 (success)
|
// Verify exit code is 0 (success)
|
||||||
@ -215,13 +217,15 @@ func TestRunWithVersionFlags(t *testing.T) {
|
|||||||
args := []string{"articulate-parser", flag}
|
args := []string{"articulate-parser", flag}
|
||||||
exitCode := run(args)
|
exitCode := run(args)
|
||||||
|
|
||||||
// Restore stdout
|
// Restore stdout. Close errors are ignored: the version output has already
|
||||||
w.Close()
|
// been written and we're about to read it; close semantics don't affect correctness.
|
||||||
|
_ = w.Close()
|
||||||
os.Stdout = oldStdout
|
os.Stdout = oldStdout
|
||||||
|
|
||||||
// Read captured output
|
// Read captured output. Copy errors are ignored: the output was successfully
|
||||||
|
// produced and we can verify its contents regardless of any I/O edge cases.
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
io.Copy(&buf, r)
|
_, _ = io.Copy(&buf, r)
|
||||||
output := buf.String()
|
output := buf.String()
|
||||||
|
|
||||||
// Verify exit code is 0 (success)
|
// Verify exit code is 0 (success)
|
||||||
@ -265,30 +269,36 @@ func TestRunWithInvalidFile(t *testing.T) {
|
|||||||
args := []string{"articulate-parser", "nonexistent-file.json", "markdown", "output.md"}
|
args := []string{"articulate-parser", "nonexistent-file.json", "markdown", "output.md"}
|
||||||
exitCode := run(args)
|
exitCode := run(args)
|
||||||
|
|
||||||
// Restore stdout/stderr and log output
|
// Restore stdout/stderr and log output. Close errors are ignored: we've already
|
||||||
stdoutW.Close()
|
// written all error messages to these pipes before closing them, and the test
|
||||||
stderrW.Close()
|
// only cares about verifying the captured output.
|
||||||
|
_ = stdoutW.Close()
|
||||||
|
_ = stderrW.Close()
|
||||||
os.Stdout = oldStdout
|
os.Stdout = oldStdout
|
||||||
os.Stderr = oldStderr
|
os.Stderr = oldStderr
|
||||||
log.SetOutput(oldLogOutput)
|
log.SetOutput(oldLogOutput)
|
||||||
|
|
||||||
// Read captured output
|
// Read captured output. Copy errors are ignored: the error messages have been
|
||||||
|
// successfully written to the pipes, and we can verify the output content
|
||||||
|
// regardless of any edge cases in pipe closure or I/O completion.
|
||||||
var stdoutBuf, stderrBuf bytes.Buffer
|
var stdoutBuf, stderrBuf bytes.Buffer
|
||||||
io.Copy(&stdoutBuf, stdoutR)
|
_, _ = io.Copy(&stdoutBuf, stdoutR)
|
||||||
io.Copy(&stderrBuf, stderrR)
|
_, _ = io.Copy(&stderrBuf, stderrR)
|
||||||
|
|
||||||
stdoutR.Close()
|
// Close read ends of pipes. Errors ignored: we've already consumed all data
|
||||||
stderrR.Close()
|
// from these pipes, and close errors don't affect test assertions.
|
||||||
|
_ = stdoutR.Close()
|
||||||
|
_ = stderrR.Close()
|
||||||
|
|
||||||
// Verify exit code
|
// Verify exit code
|
||||||
if exitCode != 1 {
|
if exitCode != 1 {
|
||||||
t.Errorf("Expected exit code 1 for non-existent file, got %d", exitCode)
|
t.Errorf("Expected exit code 1 for non-existent file, got %d", exitCode)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Should have error output
|
// Should have error output in structured log format
|
||||||
errorOutput := stderrBuf.String()
|
output := stdoutBuf.String()
|
||||||
if !strings.Contains(errorOutput, "Error processing course") {
|
if !strings.Contains(output, "level=ERROR") && !strings.Contains(output, "failed to process course") {
|
||||||
t.Errorf("Expected error message about processing course, got: %s", errorOutput)
|
t.Errorf("Expected error message about processing course, got: %s", output)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -312,30 +322,36 @@ func TestRunWithInvalidURI(t *testing.T) {
|
|||||||
args := []string{"articulate-parser", "https://example.com/invalid", "markdown", "output.md"}
|
args := []string{"articulate-parser", "https://example.com/invalid", "markdown", "output.md"}
|
||||||
exitCode := run(args)
|
exitCode := run(args)
|
||||||
|
|
||||||
// Restore stdout/stderr and log output
|
// Restore stdout/stderr and log output. Close errors are ignored: we've already
|
||||||
stdoutW.Close()
|
// written all error messages about the invalid URI to these pipes before closing,
|
||||||
stderrW.Close()
|
// and test correctness only depends on verifying the captured error output.
|
||||||
|
_ = stdoutW.Close()
|
||||||
|
_ = stderrW.Close()
|
||||||
os.Stdout = oldStdout
|
os.Stdout = oldStdout
|
||||||
os.Stderr = oldStderr
|
os.Stderr = oldStderr
|
||||||
log.SetOutput(oldLogOutput)
|
log.SetOutput(oldLogOutput)
|
||||||
|
|
||||||
// Read captured output
|
// Read captured output. Copy errors are ignored: the error messages have been
|
||||||
|
// successfully written and we can verify the failure output content regardless
|
||||||
|
// of any edge cases in pipe lifecycle or I/O synchronization.
|
||||||
var stdoutBuf, stderrBuf bytes.Buffer
|
var stdoutBuf, stderrBuf bytes.Buffer
|
||||||
io.Copy(&stdoutBuf, stdoutR)
|
_, _ = io.Copy(&stdoutBuf, stdoutR)
|
||||||
io.Copy(&stderrBuf, stderrR)
|
_, _ = io.Copy(&stderrBuf, stderrR)
|
||||||
|
|
||||||
stdoutR.Close()
|
// Close read ends of pipes. Errors ignored: we've already consumed all data
|
||||||
stderrR.Close()
|
// and close errors don't affect the validation of the error output.
|
||||||
|
_ = stdoutR.Close()
|
||||||
|
_ = stderrR.Close()
|
||||||
|
|
||||||
// Should fail because the URI is invalid/unreachable
|
// Should fail because the URI is invalid/unreachable
|
||||||
if exitCode != 1 {
|
if exitCode != 1 {
|
||||||
t.Errorf("Expected failure (exit code 1) for invalid URI, got %d", exitCode)
|
t.Errorf("Expected failure (exit code 1) for invalid URI, got %d", exitCode)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Should have error output
|
// Should have error output in structured log format
|
||||||
errorOutput := stderrBuf.String()
|
output := stdoutBuf.String()
|
||||||
if !strings.Contains(errorOutput, "Error processing course") {
|
if !strings.Contains(output, "level=ERROR") && !strings.Contains(output, "failed to process course") {
|
||||||
t.Errorf("Expected error message about processing course, got: %s", errorOutput)
|
t.Errorf("Expected error message about processing course, got: %s", output)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -365,16 +381,29 @@ func TestRunWithValidJSONFile(t *testing.T) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Failed to create temp file: %v", err)
|
t.Fatalf("Failed to create temp file: %v", err)
|
||||||
}
|
}
|
||||||
defer os.Remove(tmpFile.Name())
|
// Ensure temporary test file is cleaned up. Remove errors are ignored because
|
||||||
|
// the test has already used the file for its purpose, and cleanup failures don't
|
||||||
|
// invalidate the test results (the OS will eventually clean up temp files).
|
||||||
|
defer func() {
|
||||||
|
_ = os.Remove(tmpFile.Name())
|
||||||
|
}()
|
||||||
|
|
||||||
if _, err := tmpFile.WriteString(testContent); err != nil {
|
if _, err := tmpFile.WriteString(testContent); err != nil {
|
||||||
t.Fatalf("Failed to write test content: %v", err)
|
t.Fatalf("Failed to write test content: %v", err)
|
||||||
}
|
}
|
||||||
tmpFile.Close()
|
// Close the temporary file. Errors are ignored because we've already written
|
||||||
|
// the test content and the main test logic (loading the file) doesn't depend
|
||||||
|
// on the success of closing this file descriptor.
|
||||||
|
_ = tmpFile.Close()
|
||||||
|
|
||||||
// Test successful run with valid file
|
// Test successful run with valid file
|
||||||
outputFile := "test-output.md"
|
outputFile := "test-output.md"
|
||||||
defer os.Remove(outputFile)
|
// Ensure test output file is cleaned up. Remove errors are ignored because the
|
||||||
|
// test has already verified the export succeeded; cleanup failures don't affect
|
||||||
|
// the test assertions.
|
||||||
|
defer func() {
|
||||||
|
_ = os.Remove(outputFile)
|
||||||
|
}()
|
||||||
|
|
||||||
// Save original stdout
|
// Save original stdout
|
||||||
originalStdout := os.Stdout
|
originalStdout := os.Stdout
|
||||||
@ -387,13 +416,17 @@ func TestRunWithValidJSONFile(t *testing.T) {
|
|||||||
args := []string{"articulate-parser", tmpFile.Name(), "markdown", outputFile}
|
args := []string{"articulate-parser", tmpFile.Name(), "markdown", outputFile}
|
||||||
exitCode := run(args)
|
exitCode := run(args)
|
||||||
|
|
||||||
// Close write end and restore stdout
|
// Close write end and restore stdout. Close errors are ignored: we've already
|
||||||
w.Close()
|
// written the success message before closing, and any close error doesn't affect
|
||||||
|
// the validity of the captured output or the test assertions.
|
||||||
|
_ = w.Close()
|
||||||
os.Stdout = originalStdout
|
os.Stdout = originalStdout
|
||||||
|
|
||||||
// Read captured output
|
// Read captured output. Copy errors are ignored: the success message was
|
||||||
|
// successfully written to the pipe, and we can verify it regardless of any
|
||||||
|
// edge cases in pipe closure or I/O synchronization.
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
io.Copy(&buf, r)
|
_, _ = io.Copy(&buf, r)
|
||||||
output := buf.String()
|
output := buf.String()
|
||||||
|
|
||||||
// Verify successful execution
|
// Verify successful execution
|
||||||
@ -401,10 +434,9 @@ func TestRunWithValidJSONFile(t *testing.T) {
|
|||||||
t.Errorf("Expected successful execution (exit code 0), got %d", exitCode)
|
t.Errorf("Expected successful execution (exit code 0), got %d", exitCode)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Verify success message
|
// Verify success message in structured log format
|
||||||
expectedMsg := fmt.Sprintf("Successfully exported course to %s", outputFile)
|
if !strings.Contains(output, "level=INFO") || !strings.Contains(output, "successfully exported course") {
|
||||||
if !strings.Contains(output, expectedMsg) {
|
t.Errorf("Expected success message in output, got: %s", output)
|
||||||
t.Errorf("Expected success message '%s' in output, got: %s", expectedMsg, output)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Verify output file was created
|
// Verify output file was created
|
||||||
@ -440,17 +472,24 @@ func TestRunIntegration(t *testing.T) {
|
|||||||
args := []string{"articulate-parser", "articulate-sample.json", format.format, format.output}
|
args := []string{"articulate-parser", "articulate-sample.json", format.format, format.output}
|
||||||
exitCode := run(args)
|
exitCode := run(args)
|
||||||
|
|
||||||
// Restore stdout
|
// Restore stdout. Close errors are ignored: the export success message
|
||||||
w.Close()
|
// has already been written and we're about to read it; close semantics
|
||||||
|
// don't affect the validity of the captured output.
|
||||||
|
_ = w.Close()
|
||||||
os.Stdout = oldStdout
|
os.Stdout = oldStdout
|
||||||
|
|
||||||
// Read captured output
|
// Read captured output. Copy errors are ignored: the output was successfully
|
||||||
|
// produced and we can verify its contents regardless of any I/O edge cases.
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
io.Copy(&buf, r)
|
_, _ = io.Copy(&buf, r)
|
||||||
output := buf.String()
|
output := buf.String()
|
||||||
|
|
||||||
// Clean up test file
|
// Clean up test file. Remove errors are ignored because the test has
|
||||||
defer os.Remove(format.output)
|
// already verified the export succeeded; cleanup failures don't affect
|
||||||
|
// the test assertions.
|
||||||
|
defer func() {
|
||||||
|
_ = os.Remove(format.output)
|
||||||
|
}()
|
||||||
|
|
||||||
// Verify successful execution
|
// Verify successful execution
|
||||||
if exitCode != 0 {
|
if exitCode != 0 {
|
||||||
|
|||||||
Reference in New Issue
Block a user