Get started with EdgeOne Pages functions ai ocr in seconds.
Next.js
This project demonstrates how to use a visual recognition large language model within Pages functions to identify text in images. The underlying model is based on Tencent's HunYuan large language model, which can be easily swapped out with other large language models using environment variables.
First, run the development server:
npm run dev
# or
yarn dev
# or
pnpm dev
# or
bun dev
Open http://localhost:3000 with your browser to see the result.
You can start editing the page by modifying app/page.tsx
. The page auto-updates as you edit the file.
This project uses next/font
to automatically optimize and load Inter, a custom Google Font.
To learn more about Next.js, take a look at the following resources:
You can check out the Next.js GitHub repository - your feedback and contributions are welcome!
This project was inspired by llama-ocr. Go check them out!