ilang-compress

Compress natural language prompts into I-Lang — AI-native structured instructions. 40-65% token savings.

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "ilang-compress" with this command: npx skills add adsorgcn/ilang-compress

I-Lang Compress

An AI-native prompt compression protocol created by a Chinese developer.

Compress natural language prompts into dense structured instructions that any AI understands natively. 40-65% token savings, zero training needed.

Why I-Lang

Token is money. Every prompt you send to GPT/Claude/Gemini, you pay by token. I-Lang compresses your instructions into a fraction of the original size — AI reads it just as well, you pay less.

How to compress

When the user asks to compress a prompt, convert it to I-Lang syntax following these rules.

Syntax

Single operation: [VERB:@ENTITY|mod1=val1,mod2=val2] Pipe chain: [VERB1:@SRC]=>[VERB2]=>[VERB3:@DST] Each step receives previous output as @PREV.

Available Verbs (62)

Data I/O: READ, WRIT, DEL, LIST, COPY, MOVE, STRM, CACH, SYNC, Π Transform: Σ, Δ, φ, ∇, DEDU, ∂, CHNK, FLAT, NEST, λ, REDU, PIVT, TRNS, ENCD, DECD, ξ, ζ, EXPN, θ, FMT Analysis: ψ, CLST, SCOR, BNCH, AUDT, VALD, CNT, μ, TRND, CORR, FRCS, ANOM Generation: CREA, DRFT, PARA, EXTD, SHRT, STYL, TMPL, FILL Output: Ω, DISP, EXPT, PRNT, LOG Meta: VERS, HELP, DESC, INTR, SELF, ECHO, NOOP

Modifiers (28)

tgt, src, dst, frm, to, scp, dep, rng, whr, mch, exc, lim, off, top, bot, fmt, lng, sty, ton, len, col, row, srt, grp, typ, enc, chr, cap

Entities (14)

@R2, @COS, @GH, @DRIVE, @LOCAL, @WORKER, @CF, @SCREEN, @LOG, @NULL, @STDIN, @SRC, @DST, @PREV

Compression Guidelines

  • Output the compressed I-Lang instruction first, then a brief explanation of what each step does.
  • Use pipe chains for multi-step operations.
  • Use Greek symbols where applicable (Σ for merge, Δ for diff, φ for filter, etc.)
  • Maximize compression while preserving complete semantics.
  • If input is ambiguous, ask the user for clarification.

Examples

Input: Read the config file from GitHub and format it as JSON Output: [READ:@GH|path=config.json]=>[FMT|fmt=json] Explanation: READ fetches from GitHub, FMT converts to JSON format. Saved: 55%

Input: Filter all fatal errors from system logs Output: [φ:@LOG|whr="lvl=fatal"] Explanation: φ (filter) selects only entries matching fatal level. Saved: 55%

Input: Read all markdown files, merge them, summarize in 3 bullets, output Output: [LIST:@LOCAL|mch="*.md"]=>[Π:READ]=>[Σ|len=3]=>[Ω] Explanation: LIST finds files, Π batch-reads, Σ summarizes to 3 items, Ω outputs. Saved: 65%

Links

Author

Built by ilang-ai from China. I-Lang is open source under MIT license.

I-Lang v2.0

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

音乐生成

Generate custom music tracks (vocal or instrumental) via OhYesAI.

Registry SourceRecently Updated
General

LaTeX Compiler

Compile LaTeX documents to PDF using pdflatex, xelatex, or lualatex with template support

Registry SourceRecently Updated
General

Formal Methods

Formal verification with Lean 4, Coq, and Z3 SMT solver

Registry SourceRecently Updated
General

苏宁帮客预约服务

苏宁帮客预约服务

Registry SourceRecently Updated