Language Translation is the communication of the meaning of a source-language text by means of an equivalent target-language text.
Translation architecture shows below:
The Translation example is implemented using the component-level microservices defined in GenAIComps. The flow chart below shows the information flow between different microservices for this example.
---
config:
flowchart:
nodeSpacing: 400
rankSpacing: 100
curve: linear
themeVariables:
fontSize: 50px
---
flowchart LR
%% Colors %%
classDef blue fill:#ADD8E6,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5
classDef orange fill:#FBAA60,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5
classDef orchid fill:#C26DBC,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5
classDef invisible fill:transparent,stroke:transparent;
style Translation-MegaService stroke:#000000
%% Subgraphs %%
subgraph Translation-MegaService["Translation MegaService "]
direction LR
LLM([LLM MicroService]):::blue
end
subgraph UserInterface[" User Interface "]
direction LR
a([User Input Query]):::orchid
UI([UI server<br>]):::orchid
end
LLM_gen{{LLM Service <br>}}
GW([Translation GateWay<br>]):::orange
NG([Nginx MicroService]):::blue
%% Questions interaction
direction LR
a[User Input Query] --> UI
a[User Input Query] --> |Need Proxy Server|NG
NG --> UI
UI --> GW
GW <==> Translation-MegaService
%% Embedding service flow
direction LR
LLM <-.-> LLM_gen
This Translation use case performs Language Translation Inference across multiple platforms. Currently, we provide the example for Intel Gaudi2 and Intel Xeon Scalable Processors, and we invite contributions from other hardware vendors to expand OPEA ecosystem.
The Translation service can be effortlessly deployed on either Intel Gaudi2 or Intel Xeon Scalable Processors.
Refer to the Gaudi Guide for instructions on deploying Translation on Gaudi.
Refer to the Xeon Guide for instructions on deploying Translation on Xeon.