🚀
Codex Editor
  • Project Overview
    • Welcome to Codex
    • Codex Editor
      • Features
      • What was Project Accelerate?
        • Project Philosophy
        • Steering Committee
    • Vision
      • Powerful Tools and Simplicity
      • Streamlining the Translation Process
      • Unopinionated Assistance
    • Architecture
    • Frequently Asked Questions
    • Translator Requirements
    • Roadmap
    • Why VS Code?
    • Multimodality
  • Translator's Copilot
    • What is Translator's Copilot?
    • Information Management
      • Resource Indexing
      • Greek/Hebrew Insights
      • Chat with Resources
      • Project-Based Insights
    • Translation Assistance
      • Translation Drafting
        • Prioritizing Semantics over Structures
        • Reranked Translation Suggestions
      • Quality Checking
        • Token matching evaluation
        • Few-shot evaluation
        • Fine-tuned Bible QA model
        • Character and word n-gram evaluation
        • Simulated GAN evaluation
        • Linguistic Anomaly Detection (LAD)
      • Back Translation
    • Orchestration
      • Translation Memory
      • Multi-Agent Simulations
      • Drafting linguistic resources
    • Intelligent Functions Library
  • Development
    • Codex Basics
      • Projects
      • The Editor
      • Extensions
        • Rendered Views
        • Language Servers
        • Custom Notebooks
        • Global State
    • Experimental Repositories
Powered by GitBook
On this page
  1. Translator's Copilot
  2. Translation Assistance
  3. Quality Checking

Few-shot evaluation

Keeping things simple, and sticking with only the resources at hand, what if the evaluation stage involved the following:

  1. Take the draft translation and find the most similar translation pairs to the paired prediction

  2. Try to evaluate if any of the words seem misused

    1. Perhaps you could have a rules-based matcher that identifies any tokens in the source+prediction pair that only occur in EITHER the source OR the target in any of the similar examples

    2. It might also be worth examining whether any words seem to be mistranslated based on the gold standard examples

    3. POSaligning these examples would probably help

  3. Pass the task back to the translate bot with special notes about any seemingly mistranslated words.

Essentially: [ most similar examples ] --> [ prediction ] --> [ most similar examples ]

PreviousToken matching evaluationNextFine-tuned Bible QA model

Last updated 1 year ago