Junyan Xu
Junyan Xu
maintainer merge
CI fails because the implicitness of the function in `pairwise_of_fn` changed, apparently.
@eric-wieser This PR seems forgotten. Do you have any other comments or do you want another maintainer to take a look?
forgot to label it awaiting-review ... now the main file has been ported to mathlib4 apparently
I just looked at the definition of [slash_action](https://leanprover-community.github.io/mathlib_docs/number_theory/modular_forms/slash_actions.html#slash_action) and found that `slash_action β G α γ` is exactly same as [`β → representation γ Gᵐᵒᵖ α`](https://leanprover-community.github.io/mathlib_docs/representation_theory/basic.html#representation ). `slash_action` has weaker...
Are there plans to merge master and port this PR to Lean 4? It's listed as the first item in @joelriou 's [AIM workshop statement](http://admin.aimath.org/resources/alggeominlean/participantlist/).
At least in Leela Zero the Dirichlet noise is generated only once for each root node. The idea of is that moves with low policy prior may get very high...
IIUC the Alpaca repo doesn't contain the model (because LLaMA isn't openly available except for the leak) but contains the code to fine-tune to make output resemble ChatGPT. [Eliezer Yudkowsky](https://twitter.com/ESYudkowsky/status/1635577836525469697)...
Thanks a lot for the info! I didn't know about Alpaca-LoRA. More is going on around LLaMA than I realized! AFAIK, being an RNN, RWKV is less resource intensive than...
https://twitter.com/piesposi_to/status/1636780485597708290 seems useful to finetune ChatRWKV for Chinese instruction following.