The Translation example demonstrates the implementation of language translation using OPEA component-level microservices.
The architecture of the Translation Application is illustrated below:
The Translation example is implemented using the component-level microservices defined in GenAIComps. The flow chart below shows the information flow between different microservices for this example.
--- config: flowchart: nodeSpacing: 400 rankSpacing: 100 curve: linear themeVariables: fontSize: 50px --- flowchart LR %% Colors %% classDef blue fill:#ADD8E6,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5 classDef orange fill:#FBAA60,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5 classDef orchid fill:#C26DBC,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5 classDef invisible fill:transparent,stroke:transparent; style Translation-MegaService stroke:#000000 %% Subgraphs %% subgraph Translation-MegaService["Translation MegaService "] direction LR LLM([LLM MicroService]):::blue end subgraph UserInterface[" User Interface "] direction LR a([User Input Query]):::orchid UI([UI server<br>]):::orchid end LLM_gen{{LLM Service <br>}} GW([Translation GateWay<br>]):::orange NG([Nginx MicroService]):::blue %% Questions interaction direction LR a[User Input Query] --> UI a[User Input Query] --> |Need Proxy Server|NG NG --> UI UI --> GW GW <==> Translation-MegaService %% Embedding service flow direction LR LLM <-.-> LLM_gen This Translation use case performs Language Translation Inference across multiple platforms. Currently, we provide the example for Intel Gaudi2, Intel Xeon Scalable Processors and AMD EPYC™ Processors, and we invite contributions from other hardware vendors to expand OPEA ecosystem.
The table below lists the available deployment options and their implementation details for different hardware platforms.
| Platform | Deployment Method | Link |
|---|---|---|
| Intel Xeon | Docker compose | Deployment on Xeon |
| Intel Gaudi2 | Docker compose | Deployment on Gaudi |
| AMD EPYC | Docker compose | Deployment on AMD EPYC |
| AMD ROCm | Docker compose | Deployment on AMD Rocm |
| Deploy Method | LLM Engine | LLM Model | Hardware |
|---|---|---|---|
| Docker Compose | vLLM, TGI | haoranxu/ALMA-13B | Intel Gaudi |
| Docker Compose | vLLM, TGI | haoranxu/ALMA-13B | Intel Xeon |
| Docker Compose | vLLM, TGI | haoranxu/ALMA-13B | AMD Xeon |
| Docker Compose | vLLM, TGI | haoranxu/ALMA-13B | AMD EPYC |
| Helm Charts | vLLM, TGI | haoranxu/ALMA-13B | Intel Gaudi |
| Helm Charts | vLLM, TGI | haoranxu/ALMA-13B | Intel Xeon |
