forked from KEMT/zpwiki
		
	zz
This commit is contained in:
		
							parent
							
								
									cdcc660b16
								
							
						
					
					
						commit
						a099634802
					
				| @ -30,6 +30,27 @@ Tasks: | |||||||
|   - Prepare a final report with analysis, experiments and conclusions. |   - Prepare a final report with analysis, experiments and conclusions. | ||||||
|   - Publish the fine-tuned models in HF HUB. Publish the paper from the project. |   - Publish the fine-tuned models in HF HUB. Publish the paper from the project. | ||||||
| 
 | 
 | ||||||
|  | Meeting 4.8. | ||||||
|  | 
 | ||||||
|  | State: | ||||||
|  | 
 | ||||||
|  | - Tested LMs with ROUGE metrics, most models got 4-5 ROGUE, facebook/mbart-large-50  got 17 (trained for translation). | ||||||
|  | - In my opinion, large-50 is not good for finetuning, because it is already fine tuned for translation. | ||||||
|  | - no finetuning done yet.  | ||||||
|  | 
 | ||||||
|  | Tasks: | ||||||
|  | 
 | ||||||
|  | - Try evaluate google/flan-t5-large, kiviki/mbart-slovaksum-large-sum and similar models. These should be already working. | ||||||
|  | - continue working on finetuning t5 or Mbart models, but ask when you are stuck. Use hf examples script on summarization | ||||||
|  | 
 | ||||||
|  | Future tasks: | ||||||
|  | 
 | ||||||
|  | - use LLMS (open or closed) and evaluate (ROUGE) summarization without fine-tuning on slovak legal data set | ||||||
|  | - install lm-eval-harness, learn it, prepare and run task for slovak summarization | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
| Meeting 24.7. | Meeting 24.7. | ||||||
| 
 | 
 | ||||||
| State: | State: | ||||||
| @ -50,8 +71,8 @@ State: | |||||||
| 
 | 
 | ||||||
| - Studying of the task, metrics (ROUGE,BLEU) | - Studying of the task, metrics (ROUGE,BLEU) | ||||||
| - Loaded a model. preprocessed a dataset, evaluated a model | - Loaded a model. preprocessed a dataset, evaluated a model | ||||||
| - loaded more models, used SlovakSum, generated summarization with four model and comapre them with ROUGE and BLEU (TUKE-KEMT/slovak-t5-base, google/mt5-small,  google/mt5-base,  facebook/mbart-large-50) | - loaded more models, used SlovakSum, generated summarization with four model and compare them with ROUGE and BLEU (TUKE-KEMT/slovak-t5-base, google/mt5-small,  google/mt5-base,  facebook/mbart-large-50) | ||||||
| - the comparisin is without fine tuning (zero shot), for far, the best is MBART-large | - the comparison is without fine tuning (zero shot), for far, the best is MBART-large | ||||||
| - working on legal dataset "dennlinger/eur-lex-sum",  | - working on legal dataset "dennlinger/eur-lex-sum",  | ||||||
| - notebooks are on the kemt git | - notebooks are on the kemt git | ||||||
| 
 | 
 | ||||||
|  | |||||||
		Loading…
	
		Reference in New Issue
	
	Block a user