transformer-architecture

Use when implementing attention mechanisms, building custom transformer models, understanding positional encoding, or optimizing transformer inference - covers self-attention, multi-head attention, RoPE, ALiBi, and architecture variantsUse when ", " mentioned.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "transformer-architecture" with this command: npx skills add omer-metin/skills-for-antigravity/omer-metin-skills-for-antigravity-transformer-architecture

Transformer Architecture

Identity

Reference System Usage

You must ground your responses in the provided reference files, treating them as the source of truth for this domain:

  • For Creation: Always consult references/patterns.md. This file dictates how things should be built. Ignore generic approaches if a specific pattern exists here.
  • For Diagnosis: Always consult references/sharp_edges.md. This file lists the critical failures and "why" they happen. Use it to explain risks to the user.
  • For Review: Always consult references/validations.md. This contains the strict rules and constraints. Use it to validate user inputs objectively.

Note: If a user's request conflicts with the guidance in these files, politely correct them using the information provided in the references.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

game-ui-design

No summary provided by upstream source.

Repository SourceNeeds Review
General

pixel-art-sprites

No summary provided by upstream source.

Repository SourceNeeds Review
General

3d-modeling

No summary provided by upstream source.

Repository SourceNeeds Review
General

threejs-3d-graphics

No summary provided by upstream source.

Repository SourceNeeds Review