Skip to content

Make the README more informative with a focus on the project itself#347

Open
Ingvarstep wants to merge 9 commits intomainfrom
new_readme
Open

Make the README more informative with a focus on the project itself#347
Ingvarstep wants to merge 9 commits intomainfrom
new_readme

Conversation

@Ingvarstep
Copy link
Copy Markdown
Collaborator

Hi @urchade ,

Because the current README highlights other commercial tools and projects, it creates the following issues:

  • The README doesn't inform well about the key features and benefits of the GLiNER project;
  • It creates questions about the maintainability of the project and its open-source goals;
  • The README doesn't clearly describe the contribution guidelines.

I have crafted a new README proposal to address the issues described above and to make it more informative and better capture the GLiNER project itself.

To highlight GLiNER2 and other amazing projects from the Ecosystem, I created a dedicated section listing them.

Looking forward to your suggestions and feedback.

@tomaarsen, it would be awesome to get your review and feedback as well.

Best,
Ihor

@Ingvarstep Ingvarstep self-assigned this Apr 20, 2026
@Ingvarstep Ingvarstep requested review from tomaarsen and urchade April 20, 2026 10:40
Copy link
Copy Markdown
Collaborator

@tomaarsen tomaarsen left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure, I can give a review 🤗

Overall some nice extensions I think, and a pretty logical structure, i.e.: title, tags, relevant banner, one paragraph, and then straight into benefits followed by a very solid quick start. I would probably clean up the em dashes though. In my eyes, they read as low-effort, when clearly a lot of love has gone into this project.

Edit: I would more clearly link to the existing models, e.g.: https://huggingface.co/models?library=gliner
I realise now that I miss that a bit.

Comment thread assets/banner.png
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ooh, looks very clean

Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It looks minimalistic, but it was actually hard to find the right prompt to generate it that way ;)

Comment thread README.md Outdated
Comment thread README.md Outdated
- [Example Notebooks](https://github.com/urchade/GLiNER/tree/main/examples)
- Finetune on Colab &nbsp;[<img align="center" src="https://colab.research.google.com/assets/colab-badge.svg" />](https://colab.research.google.com/drive/1HNKd74cmfS9tGvWrKeIjSxBt01QQS7bq?usp=sharing)
## 🛠 Installation & Usage
---
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is very minor, but I think the --- lines are not needed when you also have a new header starting below each

Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actually, I looked into it a bit more, and we can definitely remove the --- lines when there is a new header after them.

Comment thread README.md
## 🛠 Installation & Usage
---

## Quick Start
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I like the updated README here, very bite-sized. All that's missing (I think) is a link to docs below each (sub)section, e.g. below the Output:, the Quanitzation and Compilation, the Serving, etc.). You already have some links below Training, that's good.

Comment thread README.md Outdated
Or apply after loading:
## Serving

GLiNER provides a built-in serving interface for batch inference:
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I would add 1 extra sentence to explain why someone would serve instead of using the script above.

Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Btw, is this merged? I see #346 which seems to refer to this

Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It's still under review, but API and documentation should be stable.

Comment thread README.md
quantize=True,
compile_torch_model=True,
)
```
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm missing the quantization/compilation gains a bit. As a user I'd like to know what this does for me.

Comment thread README.md

---

<div align="center">
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Not very crucial in my opinion, but it's fine to keep.

Comment thread README.md Outdated
- `quantize="bf16"` — bfloat16. Better numerical stability, slightly less speedup (~1.2x).
- `quantize="int8"` — int8 quantization. On CPU, uses built-in FBGEMM int8 kernels (~1.6x speedup). On GPU, uses [torchao](https://github.com/pytorch/ao) int8 weight-only quantization (~50% memory reduction, no speed gain). Intended for models fine-tuned with quantization-aware training (QAT). Stock DeBERTa-based models lose accuracy with int8.
- On CPU, fp16/bf16 quantization reduces memory usage but does not improve speed.
| Architecture | Description |
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice to see this, I honestly lost track and I've been following GLiNER-related developments pretty close

Comment thread README.md Outdated
| **Uni-encoder** | Strong zero-shot capabilities, supports up to ~50 entity types. The original GLiNER architecture. |
| **Bi-encoder** | Scalable to massive numbers of entity types via separate text and label encoding. |
| **RelEx** | Joint NER and relation extraction in a single model. |
| **GLiNER Decoder** | Hybrid architecture for open NER — entity types are generated with a small decoder for maximum flexibility. |
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
| **GLiNER Decoder** | Hybrid architecture for open NER entity types are generated with a small decoder for maximum flexibility. |
| **GLiNER Decoder** | Hybrid architecture for open NER: entity types are generated with a small decoder for maximum flexibility. |

Personal preference: I would remove the em dashes throughout

Comment thread README.md

---

## Popular Use Cases
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Even better if you can immediately link from a use case to a matching example like these: https://urchade.github.io/GLiNER/usage.html#example-1-extract-company-information

Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is a good point, I am adding cross-links to the documentation with real examples.

@Ingvarstep
Copy link
Copy Markdown
Collaborator Author

Hi @urchade , any thoughts on the readme update?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants