Skip to content

Commit

Permalink
Merge branch 'main' into release_notes_v1_1
Browse files Browse the repository at this point in the history
  • Loading branch information
ftian1 authored Nov 26, 2024
2 parents c16ba07 + dc92900 commit 3896d06
Show file tree
Hide file tree
Showing 9 changed files with 456 additions and 17 deletions.
2 changes: 2 additions & 0 deletions conf.py
Original file line number Diff line number Diff line change
Expand Up @@ -70,6 +70,8 @@
# files and directories to ignore when looking for source files.
exclude_patterns = [
'scripts/*',
'examples/AgentQnA/deploy/index.rst',
'examples/AgentQnA/deploy/xeon.md'
]
try:
import sphinx_rtd_theme
Expand Down
9 changes: 8 additions & 1 deletion examples/AgentQnA/AgentQnA_Guide.rst
Original file line number Diff line number Diff line change
Expand Up @@ -43,5 +43,12 @@ The worker agent uses the retrieval tool to generate answers to the queries post

Deployment
**********
Here are some deployment options, depending on your hardware and environment:

See the :ref:`agentqna-example-deployment`.
Single Node
+++++++++++++++
.. toctree::
:maxdepth: 1

Xeon Scalable Processor </GenAIExamples/AgentQnA/docker_compose/intel/cpu/xeon/README.md>
Gaudi </GenAIExamples/AgentQnA/docker_compose/intel/hpu/gaudi/README.md>
7 changes: 2 additions & 5 deletions examples/AgentQnA/deploy/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,6 @@ AgentQnA Example Deployment Options
Here are some deployment options, depending on your hardware and environment:

Single Node
***********

.. toctree::
:maxdepth: 1

Xeon Scalable Processor <xeon>
- **Xeon Scalable Processor**: `Xeon <https://opea-project.github.io/latest/GenAIExamples/AgentQnA/docker_compose/intel/cpu/xeon/README.html>`_
- **Gaudi**: `Gaudi <https://opea-project.github.io/latest/GenAIExamples/AgentQnA/docker_compose/intel/hpu/gaudi/README.html>`_
14 changes: 7 additions & 7 deletions examples/ChatQnA/ChatQnA_Guide.rst
Original file line number Diff line number Diff line change
Expand Up @@ -204,16 +204,16 @@ The gateway serves as the interface for users to access. The gateway routes inco
Deployment
**********

See the :ref:`chatqna-example-deployment` that includes both single-node and
Here are some deployment options, including both single-node and
orchestrated multi-node configurations, and choose the one that best fits your
requirements. Here are quick references to the single-node deployment options:

* :doc:`Xeon Scalable Processor <deploy/xeon>`
* :doc:`Gaudi AI Accelerator <deploy/gaudi>`
* :doc:`Nvidia GPU <deploy/nvidia>`
* :doc:`AI PC <deploy/aipc>`
requirements.

.. toctree::
:maxdepth: 1

ChatQnA Deployment Options <deploy/index>

----

Troubleshooting
***************
Expand Down
4 changes: 2 additions & 2 deletions examples/ChatQnA/deploy/gaudi.md
Original file line number Diff line number Diff line change
Expand Up @@ -416,7 +416,7 @@ CONTAINER ID IMAGE COMMAND
ce4e7802a371 opea/retriever-redis:${TAG} "python retriever_re…" About a minute ago Up About a minute 0.0.0.0:7000->7000/tcp, :::7000->7000/tcp retriever-redis-server
be6cd2d0ea38 opea/reranking-tei:${TAG} "python reranking_te…" About a minute ago Up About a minute 0.0.0.0:8000->8000/tcp, :::8000->8000/tcp reranking-tei-gaudi-server
cc45ff032e8c opea/tei-gaudi:${TAG} "text-embeddings-rou…" About a minute ago Up About a minute 0.0.0.0:8090->80/tcp, :::8090->80/tcp tei-embedding-gaudi-server
4969ec3aea02 opea/llm-vllm-hpu:${TAG} "/bin/bash -c 'expor…" About a minute ago Up About a minute 0.0.0.0:8007->80/tcp, :::8007->80/tcp vllm-gaudi-server
4969ec3aea02 opea/vllm-gaudi:${TAG} "/bin/bash -c 'expor…" About a minute ago Up About a minute 0.0.0.0:8007->80/tcp, :::8007->80/tcp vllm-gaudi-server
0657cb66df78 redis/redis-stack:7.2.0-v9 "/entrypoint.sh" About a minute ago Up About a minute 0.0.0.0:6379->6379/tcp, :::6379->6379/tcp, 0.0.0.0:8001->8001/tcp, :::8001->8001/tcp redis-vector-db
684d3e9d204a ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 "text-embeddings-rou…" About a minute ago Up About a minute 0.0.0.0:8808->80/tcp, :::8808->80/tcp tei-reranking-gaudi-server
```
Expand Down Expand Up @@ -863,7 +863,7 @@ View the docker input parameters in `./ChatQnA/docker_compose/intel/hpu/gaudi/c

```yaml
vllm-service:
image: ${REGISTRY:-opea}/llm-vllm-hpu:${TAG:-latest}
image: ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
container_name: vllm-gaudi-server
ports:
- "8007:80"
Expand Down
45 changes: 45 additions & 0 deletions examples/CodeGen/CodeGen_Guide.rst
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
.. _Codegen_Guide:

Codegen Sample Guide
#####################

.. note:: This guide is in its early development and is a work-in-progress with
placeholder content.

Overview
********

The CodeGen example uses specialized AI models that went through training with datasets that
encompass repositories, documentation, programming code, and web data. With an understanding
of various programming languages, coding patterns, and software development concepts, the
CodeGen LLMs assist developers and programmers. The LLMs can be integrated into the developers'
Integrated Development Environments (IDEs) to have more contextual awareness to write more
refined and relevant code based on the suggestions.

Purpose
*******
* Code Generation: Streamline coding through Code Generation, enabling non-programmers to describe tasks for code creation.
* Code Completion: Accelerate coding by suggesting contextually relevant snippets as developers type.
* Code Translation and Modernization: Translate and modernize code across multiple programming languages, aiding interoperability and updating legacy projects.
* Code Summarization: Extract key insights from codebases, improving readability and developer productivity.
* Code Refactoring: Offer suggestions for code refactoring, enhancing code performance and efficiency.
* AI-Assisted Testing: Assist in creating test cases, ensuring code robustness and accelerating development cycles.
* Error Detection and Debugging: Detect errors in code and provide detailed descriptions and potential fixes, expediting debugging processes.

How It Works
************

The CodeGen example uses an open-source code generation model with Text Generation Inference (TGI)
for serving deployment. It is presented as a Code Copilot application as shown in the diagram below.

.. figure:: /GenAIExamples/CodeGen/assets/img/codegen_architecture.png
:alt: CodeGen Architecture Diagram

Deployment
**********
Here are some deployment options, depending on your hardware and environment:

.. toctree::
:maxdepth: 1

CodeGen Deployment Options <deploy/index>
Loading

0 comments on commit 3896d06

Please sign in to comment.