Spaces:
Sleeping
Sleeping
| { | |
| "users": [], | |
| "learning_sessions": { | |
| "default": { | |
| "position": { | |
| "x": 13, | |
| "y": 7 | |
| }, | |
| "level": 5, | |
| "totalReward": 900, | |
| "visitedResources": [ | |
| "18", | |
| "13", | |
| "3", | |
| "16", | |
| "17", | |
| "1", | |
| "10", | |
| "8", | |
| "11", | |
| "6", | |
| "15", | |
| "2", | |
| "7", | |
| "4", | |
| "5", | |
| "14", | |
| "12", | |
| "9" | |
| ] | |
| } | |
| }, | |
| "polylines": { | |
| "polyline_0": { | |
| "id": "polyline_0", | |
| "name": "hi", | |
| "path": [], | |
| "color": "rgba(124, 111, 197, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "hiiii" | |
| }, | |
| "polyline_1": { | |
| "id": "polyline_1", | |
| "name": "HI", | |
| "path": [], | |
| "color": "rgba(233, 165, 150, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "HII" | |
| }, | |
| "polyline_2": { | |
| "id": "polyline_2", | |
| "name": "Hey", | |
| "path": [], | |
| "color": "rgba(209, 233, 75, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "Helloo", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.5484583564915045, | |
| 0.46220741047368424, | |
| 0.46279993894614624, | |
| 0.4622026423083014, | |
| 0.48167479720554324, | |
| 0.5185239739531438, | |
| 0.47663528950961, | |
| 0.5005995736735275, | |
| 0.45502656055000007, | |
| 0.4580010883357644, | |
| 0.45637659612686454, | |
| 0.5112952634097881, | |
| 0.5216384772372223, | |
| 0.5355484607706302, | |
| 0.5135652915368838, | |
| 0.46179487959519183, | |
| 0.5207722130936243, | |
| 0.49757484056057777 | |
| ] | |
| }, | |
| "polyline_3": { | |
| "id": "polyline_3", | |
| "name": "finetuning", | |
| "path": [], | |
| "color": "rgba(121, 211, 92, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "I studied fine tuning and RAG", | |
| "keywords_found": [ | |
| "Fine tuning LLM", | |
| "Retrieval Augmented Generation" | |
| ], | |
| "module_scores": [ | |
| 0.49860121134816787, | |
| 0.4500395442873277, | |
| 0.471618804424644, | |
| 0.6583558195456436, | |
| 0.4580364051366949, | |
| 0.5067814340383607, | |
| 0.5267787555517994, | |
| 0.45159593526792297, | |
| 0.5354930330453695, | |
| 0.4633757771977945, | |
| 0.7440110046152995, | |
| 0.5165231405422576, | |
| 0.5131620206629675, | |
| 0.5382607288711271, | |
| 0.548322797358488, | |
| 0.4885586287263669, | |
| 0.5297033776671244, | |
| 0.5269894013550447 | |
| ] | |
| }, | |
| "polyline_4": { | |
| "id": "polyline_4", | |
| "name": "main", | |
| "path": [], | |
| "color": "rgba(213, 61, 92, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "I studied fine tuning and RAG", | |
| "keywords_found": [ | |
| "Fine tuning LLM", | |
| "Retrieval Augmented Generation" | |
| ], | |
| "module_scores": [ | |
| 0.15775318443775177, | |
| 0.20897206664085388, | |
| 0.20236347615718842, | |
| 0.6945049047470093, | |
| 0.4463333487510681, | |
| 0.12979882955551147, | |
| 0.4455224275588989, | |
| 0.1599089652299881, | |
| 0.10129339247941971, | |
| 0.19735002517700195, | |
| 0.412064354121685, | |
| 0.15838992595672607, | |
| 0.1395527720451355, | |
| 0.06164519488811493, | |
| 0.19151057302951813, | |
| 0.05123066157102585, | |
| 0.046856656670570374, | |
| 0.09634523838758469 | |
| ] | |
| }, | |
| "polyline_5": { | |
| "id": "polyline_5", | |
| "name": "Test Summary", | |
| "path": [ | |
| { | |
| "x": 7, | |
| "y": 0 | |
| }, | |
| { | |
| "x": 14, | |
| "y": 1 | |
| } | |
| ], | |
| "color": "rgba(63, 91, 199, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.73, | |
| "summary": "I learned about BERT and Transformers.", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.41952282190322876, | |
| 0.4550943672657013, | |
| 0.38158124685287476, | |
| 0.22420939803123474, | |
| 0.21500593423843384, | |
| 0.2883783280849457, | |
| 0.08223940432071686, | |
| 0.2931893765926361, | |
| 0.12915080785751343, | |
| 0.09058266133069992, | |
| 0.19895221292972565, | |
| 0.10976134240627289, | |
| 0.10291370004415512, | |
| 0.17602239549160004, | |
| 0.28020530939102173, | |
| 0.2765440344810486, | |
| 0.153752401471138, | |
| 0.1386902928352356 | |
| ] | |
| }, | |
| "polyline_6": { | |
| "id": "polyline_6", | |
| "name": "work", | |
| "path": [ | |
| { | |
| "x": 9, | |
| "y": 18 | |
| }, | |
| { | |
| "x": 1, | |
| "y": 15 | |
| }, | |
| { | |
| "x": 6, | |
| "y": 19 | |
| }, | |
| { | |
| "x": 7, | |
| "y": 19 | |
| } | |
| ], | |
| "color": "rgba(181, 126, 100, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.76, | |
| "summary": "I have learned a lot about rag and preprocessing it was pretty good understanding.", | |
| "keywords_found": [ | |
| "Retrieval Augmented Generation" | |
| ], | |
| "module_scores": [ | |
| 0.2574118673801422, | |
| 0.2806059718132019, | |
| 0.18621119856834412, | |
| 0.10942331701517105, | |
| 0.26856184005737305, | |
| 0.3550261855125427, | |
| 0.07388029992580414, | |
| 0.3707934617996216, | |
| 0.1708703488111496, | |
| 0.2050769031047821, | |
| 0.5013039708137512, | |
| 0.1576768010854721, | |
| 0.10632561147212982, | |
| 0.12748579680919647, | |
| 0.19742925465106964, | |
| 0.05362231656908989, | |
| 0.07185838371515274, | |
| 0.055319223552942276 | |
| ] | |
| }, | |
| "polyline_7": { | |
| "id": "polyline_7", | |
| "name": "Hey", | |
| "path": [], | |
| "color": "rgba(102, 156, 166, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.017389550805091858, | |
| 0.10327564924955368, | |
| 0.09360406547784805, | |
| 0.03690618276596069, | |
| 0.04519070312380791, | |
| 0.038354936987161636, | |
| 0.07477933913469315, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 0.002015892416238785, | |
| 0.07473935186862946, | |
| 0.0972217470407486, | |
| 0.0, | |
| 0.05581950396299362 | |
| ] | |
| }, | |
| "polyline_8": { | |
| "id": "polyline_8", | |
| "name": "DQN Learning", | |
| "path": [ | |
| { | |
| "x": 5, | |
| "y": 2 | |
| }, | |
| { | |
| "x": 9, | |
| "y": 10 | |
| } | |
| ], | |
| "color": "rgba(228, 246, 240, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.73, | |
| "summary": "I studied Policy learning using DQN and reinforcement learning agents.", | |
| "keywords_found": [ | |
| "Agentic AI", | |
| "Policy learning using DQN" | |
| ], | |
| "module_scores": [ | |
| 0.3080591907103858, | |
| 0.32459903260072087, | |
| 0.10327564924955368, | |
| 0.3576170255740485, | |
| 0.03690618276596069, | |
| 0.5689638306697211, | |
| 0.038354936987161636, | |
| 0.7058747758467994, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 1.0, | |
| 0.07473935186862946, | |
| 0.0972217470407486, | |
| 1.0, | |
| 1.0 | |
| ] | |
| }, | |
| "polyline_9": { | |
| "id": "polyline_9", | |
| "name": "heyyy", | |
| "path": [], | |
| "color": "rgba(144, 174, 69, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.3080591907103858, | |
| 0.32459903260072087, | |
| 0.6197177293813891, | |
| 0.5326734754360385, | |
| 0.03690618276596069, | |
| 0.5689638306697211, | |
| 0.038354936987161636, | |
| 0.7058747758467994, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.15920396625167793, | |
| 1.0, | |
| 0.36289105295307106, | |
| 0.5652326095021434, | |
| 1.0, | |
| 1.0 | |
| ] | |
| }, | |
| "polyline_10": { | |
| "id": "polyline_10", | |
| "name": "hii", | |
| "path": [], | |
| "color": "rgba(78, 166, 73, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "hiii", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.5118434429168701, | |
| 0.7275236248970032, | |
| 0.5195397734642029, | |
| 0.2506510019302368, | |
| 0.5906199216842651, | |
| 0.460385799407959, | |
| 0.9057682156562805, | |
| 0.577440083026886, | |
| 0.44362306594848633, | |
| 0.6044198870658875, | |
| 0.26611554622650146, | |
| 0.5114411115646362, | |
| 0.49167001247406006, | |
| 0.6330015063285828, | |
| 0.38416922092437744, | |
| 0.5338788628578186, | |
| 0.37343931198120117, | |
| 0.7777066826820374 | |
| ] | |
| }, | |
| "polyline_11": { | |
| "id": "polyline_11", | |
| "name": "hii", | |
| "path": [], | |
| "color": "rgba(199, 83, 212, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "nocajco", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.035517044365406036, | |
| 0.07075861096382141, | |
| 0.021871453151106834, | |
| 0.00989073608070612, | |
| 0.019412130117416382, | |
| 0.0, | |
| 0.11922578513622284, | |
| 0.039588794112205505, | |
| 0.06686414033174515, | |
| 0.09156952798366547, | |
| 0.03569290041923523, | |
| 0.0679011419415474, | |
| 0.03752126544713974, | |
| 0.00478791818022728, | |
| 0.07304392009973526, | |
| 0.011222838424146175, | |
| 0.026982085779309273, | |
| 0.054899897426366806 | |
| ] | |
| }, | |
| "polyline_12": { | |
| "id": "polyline_12", | |
| "name": "hii", | |
| "path": [], | |
| "color": "rgba(159, 210, 119, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "heyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.03199615329504013, | |
| 0.11629796773195267, | |
| 0.09658470749855042, | |
| 0.03200055658817291, | |
| 0.054899267852306366, | |
| 0.03489005193114281, | |
| 0.10348626226186752, | |
| 0.008831196464598179, | |
| 0.0, | |
| 0.0008121263235807419, | |
| 0.0, | |
| 0.051078762859106064, | |
| 0.01544315367937088, | |
| 0.07961554080247879, | |
| 0.11205136775970459, | |
| 0.007900607772171497, | |
| 0.0693073719739914 | |
| ] | |
| }, | |
| "polyline_13": { | |
| "id": "polyline_13", | |
| "name": "hello", | |
| "path": [ | |
| { | |
| "x": 3, | |
| "y": 19 | |
| } | |
| ], | |
| "color": "rgba(230, 228, 186, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "rag and everything was good", | |
| "keywords_found": [ | |
| "Retrieval Augmented Generation" | |
| ], | |
| "module_scores": [ | |
| 0.05804816260933876, | |
| 0.09254653751850128, | |
| 0.10877598077058792, | |
| 0.012610942125320435, | |
| 0.03797341510653496, | |
| 0.015777718275785446, | |
| 0.059225354343652725, | |
| 0.028679510578513145, | |
| 0.055405281484127045, | |
| 0.10398496687412262, | |
| 0.31428444385528564, | |
| 0.0, | |
| 0.0, | |
| 0.002234384650364518, | |
| 0.0014527571620419621, | |
| 0.025889771059155464, | |
| 0.012727380730211735, | |
| 0.0 | |
| ] | |
| }, | |
| "polyline_14": { | |
| "id": "polyline_14", | |
| "name": "hello", | |
| "path": [ | |
| { | |
| "x": 13, | |
| "y": 15 | |
| } | |
| ], | |
| "color": "rgba(53, 84, 159, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "rag and many other things such as pretrainig objectives and all were good", | |
| "keywords_found": [ | |
| "Pre training objectives", | |
| "Retrieval Augmented Generation" | |
| ], | |
| "module_scores": [ | |
| 0.8379164934158325, | |
| 0.3340357542037964, | |
| 0.17760807275772095, | |
| 0.22517074644565582, | |
| 0.26595813035964966, | |
| 0.23630976676940918, | |
| 0.28637999296188354, | |
| 0.19740065932273865, | |
| 0.15292729437351227, | |
| 0.13514529168605804, | |
| 0.42122960090637207, | |
| 0.10186541825532913, | |
| 0.11643099039793015, | |
| 0.16304181516170502, | |
| 0.1428326964378357, | |
| 0.1396074891090393, | |
| 0.21233661472797394, | |
| 0.17357251048088074 | |
| ] | |
| }, | |
| "polyline_15": { | |
| "id": "polyline_15", | |
| "name": "hey ", | |
| "path": [], | |
| "color": "rgba(167, 117, 238, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "hey", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.023556426167488098, | |
| 0.07454296201467514, | |
| 0.10734860599040985, | |
| 0.03490018844604492, | |
| 0.061513230204582214, | |
| 0.06893599033355713, | |
| 0.11837682127952576, | |
| 0.01902083493769169, | |
| 0.02021801471710205, | |
| 0.020470205694437027, | |
| 0.0, | |
| 0.05460038036108017, | |
| 0.018192946910858154, | |
| 0.034782107919454575, | |
| 0.10332289338111877, | |
| 0.011397147551178932, | |
| 0.07992972433567047 | |
| ] | |
| }, | |
| "polyline_16": { | |
| "id": "polyline_16", | |
| "name": "hey", | |
| "path": [], | |
| "color": "rgba(167, 64, 198, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "hey", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.023556426167488098, | |
| 0.07454296201467514, | |
| 0.10734860599040985, | |
| 0.03490018844604492, | |
| 0.061513230204582214, | |
| 0.06893599033355713, | |
| 0.11837682127952576, | |
| 0.01902083493769169, | |
| 0.02021801471710205, | |
| 0.020470205694437027, | |
| 0.0, | |
| 0.05460038036108017, | |
| 0.018192946910858154, | |
| 0.034782107919454575, | |
| 0.10332289338111877, | |
| 0.011397147551178932, | |
| 0.07992972433567047 | |
| ] | |
| }, | |
| "polyline_17": { | |
| "id": "polyline_17", | |
| "name": "hey", | |
| "path": [], | |
| "color": "rgba(142, 153, 121, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "hey", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.023556426167488098, | |
| 0.07454296201467514, | |
| 0.10734860599040985, | |
| 0.03490018844604492, | |
| 0.061513230204582214, | |
| 0.06893599033355713, | |
| 0.11837682127952576, | |
| 0.01902083493769169, | |
| 0.02021801471710205, | |
| 0.020470205694437027, | |
| 0.0, | |
| 0.05460038036108017, | |
| 0.018192946910858154, | |
| 0.034782107919454575, | |
| 0.10332289338111877, | |
| 0.011397147551178932, | |
| 0.07992972433567047 | |
| ] | |
| }, | |
| "polyline_18": { | |
| "id": "polyline_18", | |
| "name": "heyy", | |
| "path": [], | |
| "color": "rgba(205, 206, 221, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.017389550805091858, | |
| 0.10327564924955368, | |
| 0.09360406547784805, | |
| 0.03690618276596069, | |
| 0.04519070312380791, | |
| 0.038354936987161636, | |
| 0.07477933913469315, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 0.002015892416238785, | |
| 0.07473935186862946, | |
| 0.0972217470407486, | |
| 0.0, | |
| 0.05581950396299362 | |
| ] | |
| }, | |
| "polyline_19": { | |
| "id": "polyline_19", | |
| "name": "heyy", | |
| "path": [], | |
| "color": "rgba(188, 202, 118, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "I love rag with my pre trining", | |
| "keywords_found": [ | |
| "Retrieval Augmented Generation" | |
| ], | |
| "module_scores": [ | |
| 0.09309744089841843, | |
| 0.12534257769584656, | |
| 0.18787196278572083, | |
| 0.05221807584166527, | |
| 0.053495265543460846, | |
| 0.14688314497470856, | |
| 0.09295167028903961, | |
| 0.13856445252895355, | |
| 0.11419141292572021, | |
| 0.0, | |
| 0.3, | |
| 0.04010733589529991, | |
| 0.0, | |
| 0.045714717358350754, | |
| 0.06977400183677673, | |
| 0.03902996703982353, | |
| 0.010432184673845768, | |
| 0.014520765282213688 | |
| ] | |
| }, | |
| "polyline_20": { | |
| "id": "polyline_20", | |
| "name": "huu", | |
| "path": [], | |
| "color": "rgba(132, 67, 193, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "huuu", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.06137967109680176, | |
| 0.025739794597029686, | |
| 0.05682278797030449, | |
| 0.0015102589968591928, | |
| 0.0, | |
| 0.0, | |
| 0.0828077420592308, | |
| 0.03640921786427498, | |
| 0.0, | |
| 0.048738643527030945, | |
| 0.0, | |
| 0.07323087751865387, | |
| 0.0, | |
| 0.06557325273752213, | |
| 0.03551648184657097, | |
| 0.0, | |
| 0.007635355927050114, | |
| 0.030927538871765137 | |
| ] | |
| }, | |
| "polyline_21": { | |
| "id": "polyline_21", | |
| "name": "hello", | |
| "path": [], | |
| "color": "rgba(171, 185, 237, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "hello", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.09007163345813751, | |
| 0.14346976578235626, | |
| 0.09423135966062546, | |
| 0.039209116250276566, | |
| 0.09866251051425934, | |
| 0.07642864435911179, | |
| 0.050438638776540756, | |
| 0.11424312740564346, | |
| 0.06147314980626106, | |
| 0.025780409574508667, | |
| 0.04125567153096199, | |
| 0.06409071385860443, | |
| 0.0545211136341095, | |
| 0.08595173060894012, | |
| 0.03564179316163063, | |
| 0.09568523615598679, | |
| 0.05039823427796364, | |
| 0.07651611417531967 | |
| ] | |
| }, | |
| "polyline_22": { | |
| "id": "polyline_22", | |
| "name": "Introduction", | |
| "path": [ | |
| { | |
| "x": 4, | |
| "y": 19 | |
| } | |
| ], | |
| "color": "rgba(117, 80, 114, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "hello", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.09007163345813751, | |
| 0.14346976578235626, | |
| 0.19423136115074158, | |
| 0.039209116250276566, | |
| 0.09866251051425934, | |
| 0.07642864435911179, | |
| 0.050438638776540756, | |
| 0.11424312740564346, | |
| 0.06147314980626106, | |
| 0.025780409574508667, | |
| 0.04125567153096199, | |
| 0.06409071385860443, | |
| 0.0545211136341095, | |
| 0.08595173060894012, | |
| 0.03564179316163063, | |
| 0.09568523615598679, | |
| 0.05039823427796364, | |
| 0.07651611417531967 | |
| ] | |
| }, | |
| "polyline_23": { | |
| "id": "polyline_23", | |
| "name": "heyyy", | |
| "path": [], | |
| "color": "rgba(76, 80, 77, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.017389550805091858, | |
| 0.10327564924955368, | |
| 0.09360406547784805, | |
| 0.03690618276596069, | |
| 0.04519070312380791, | |
| 0.038354936987161636, | |
| 0.07477933913469315, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 0.002015892416238785, | |
| 0.07473935186862946, | |
| 0.0972217470407486, | |
| 0.0, | |
| 0.05581950396299362 | |
| ] | |
| }, | |
| "polyline_24": { | |
| "id": "polyline_24", | |
| "name": "heyyy", | |
| "path": [], | |
| "color": "rgba(123, 83, 128, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "heyyyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.015168394893407822, | |
| 0.09509415924549103, | |
| 0.08808448165655136, | |
| 0.055125169456005096, | |
| 0.03878949210047722, | |
| 0.06340508162975311, | |
| 0.0406477153301239, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.012255707755684853, | |
| 0.07024804502725601, | |
| 0.005069136619567871, | |
| 0.06734798103570938, | |
| 0.08735782653093338, | |
| 0.0, | |
| 0.03643130138516426 | |
| ] | |
| }, | |
| "polyline_25": { | |
| "id": "polyline_25", | |
| "name": "Introduction to transformers", | |
| "path": [ | |
| { | |
| "x": 4, | |
| "y": 19 | |
| } | |
| ], | |
| "color": "rgba(143, 130, 92, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "Transformers are a class of deep learning models designed to process sequential data efficiently by relying on an attention-based mechanism rather than recurrence or convolution. Introduced in the landmark paper \u201cAttention Is All You Need\u201d, transformers revolutionized natural language processing by enabling models to capture long-range dependencies in data with high parallelism.\n\nAt the core of a transformer is the self-attention mechanism, which allows the model to weigh the importance of different parts of an input sequence when processing each element. This makes transformers highly effective at understanding context, relationships, and structure within data. The architecture typically consists of an encoder\u2013decoder structure, where encoders extract meaningful representations from input data and decoders generate output sequences based on those representations.\n\nTransformers also use positional encoding to retain information about the order of tokens, since the model itself does not process data sequentially. Combined with multi-head attention and feed-forward neural networks, this design enables scalable training on large datasets.\n\nDue to their flexibility and performance, transformers form the backbone of many modern AI systems, including large language models, machine translation systems, text summarization tools, and increasingly, applications in vision, speech, and multimodal learning.", | |
| "keywords_found": [ | |
| "Multimodal LLMs" | |
| ], | |
| "module_scores": [ | |
| 0.433811753988266, | |
| 0.4582192003726959, | |
| 0.5164702534675598, | |
| 0.2946617603302002, | |
| 0.37847253680229187, | |
| 0.2834630310535431, | |
| 0.2269679754972458, | |
| 0.23173178732395172, | |
| 0.12193305790424347, | |
| 0.15064716339111328, | |
| 0.20222218334674835, | |
| 0.23586700856685638, | |
| 0.2287927120923996, | |
| 0.2737714946269989, | |
| 0.7030925154685974, | |
| 0.5322507619857788, | |
| 0.2212483286857605, | |
| 0.202763170003891 | |
| ] | |
| }, | |
| "polyline_26": { | |
| "id": "polyline_26", | |
| "name": "heyy", | |
| "path": [ | |
| { | |
| "x": 4, | |
| "y": 19 | |
| }, | |
| { | |
| "x": 7, | |
| "y": 18 | |
| }, | |
| { | |
| "x": 18, | |
| "y": 6 | |
| }, | |
| { | |
| "x": 9, | |
| "y": 15 | |
| }, | |
| { | |
| "x": 9, | |
| "y": 17 | |
| } | |
| ], | |
| "color": "rgba(94, 122, 221, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.77, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.017389550805091858, | |
| 0.2032756507396698, | |
| 0.19360406696796417, | |
| 0.13690617680549622, | |
| 0.04519070312380791, | |
| 0.13835494220256805, | |
| 0.07477933913469315, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 0.002015892416238785, | |
| 0.17473936080932617, | |
| 0.0972217470407486, | |
| 0.0, | |
| 0.05581950396299362 | |
| ] | |
| }, | |
| "polyline_27": { | |
| "id": "polyline_27", | |
| "name": "heyy", | |
| "path": [ | |
| { | |
| "x": 4, | |
| "y": 19 | |
| } | |
| ], | |
| "color": "rgba(74, 171, 224, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.017389550805091858, | |
| 0.2032756507396698, | |
| 0.09360406547784805, | |
| 0.03690618276596069, | |
| 0.04519070312380791, | |
| 0.038354936987161636, | |
| 0.07477933913469315, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 0.002015892416238785, | |
| 0.07473935186862946, | |
| 0.0972217470407486, | |
| 0.0, | |
| 0.05581950396299362 | |
| ] | |
| }, | |
| "polyline_28": { | |
| "id": "polyline_28", | |
| "name": "introduction to my ", | |
| "path": [ | |
| { | |
| "x": 4, | |
| "y": 19 | |
| } | |
| ], | |
| "color": "rgba(76, 227, 191, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "transformers", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.024954048916697502, | |
| 0.13416121900081635, | |
| 0.5420874357223511, | |
| 0.07039379328489304, | |
| 0.021926987916231155, | |
| 0.0, | |
| 0.19421246647834778, | |
| 0.0, | |
| 0.0, | |
| 0.0374816469848156, | |
| 0.0, | |
| 0.057812318205833435, | |
| 0.04283241927623749, | |
| 0.0, | |
| 0.061064332723617554, | |
| 0.2036512792110443, | |
| 0.0, | |
| 0.0 | |
| ] | |
| }, | |
| "polyline_29": { | |
| "id": "polyline_29", | |
| "name": "hiii", | |
| "path": [ | |
| { | |
| "x": 4, | |
| "y": 19 | |
| }, | |
| { | |
| "x": 7, | |
| "y": 18 | |
| } | |
| ], | |
| "color": "rgba(253, 134, 192, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.73, | |
| "summary": "transformers", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.024954048916697502, | |
| 0.13416121900081635, | |
| 0.5420874357223511, | |
| 0.17039379477500916, | |
| 0.021926987916231155, | |
| 0.0, | |
| 0.19421246647834778, | |
| 0.0, | |
| 0.0, | |
| 0.0374816469848156, | |
| 0.0, | |
| 0.057812318205833435, | |
| 0.04283241927623749, | |
| 0.0, | |
| 0.061064332723617554, | |
| 0.2036512792110443, | |
| 0.0, | |
| 0.0 | |
| ] | |
| }, | |
| "polyline_30": { | |
| "id": "polyline_30", | |
| "name": "heyy", | |
| "path": [ | |
| { | |
| "x": 4, | |
| "y": 15 | |
| }, | |
| { | |
| "x": 11, | |
| "y": 19 | |
| }, | |
| { | |
| "x": 5, | |
| "y": 14 | |
| } | |
| ], | |
| "color": "rgba(109, 71, 172, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.74, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.017389550805091858, | |
| 0.10327564924955368, | |
| 0.09360406547784805, | |
| 0.13690617680549622, | |
| 0.04519070312380791, | |
| 0.13835494220256805, | |
| 0.07477933913469315, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 0.002015892416238785, | |
| 0.17473936080932617, | |
| 0.0972217470407486, | |
| 0.0, | |
| 0.05581950396299362 | |
| ], | |
| "strengths": [ | |
| "Instruction tuning", | |
| "Parameter efficient fine tuning", | |
| "Multimodal LLMs" | |
| ], | |
| "dominant_topics": [], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with . You successfully reinforced concepts in Instruction tuning, Parameter efficient fine tuning. Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_31": { | |
| "id": "polyline_31", | |
| "name": "Introduction to transformers", | |
| "path": [], | |
| "color": "rgba(112, 160, 225, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "Transformers are a class of deep learning models designed to process sequential data efficiently by relying on an attention-based mechanism rather than recurrence or convolution. Introduced in the landmark paper \u201cAttention Is All You Need\u201d, transformers revolutionized natural language processing by enabling models to capture long-range dependencies in data with high parallelism.\n\nAt the core of a transformer is the self-attention mechanism, which allows the model to weigh the importance of different parts of an input sequence when processing each element. This makes transformers highly effective at understanding context, relationships, and structure within data. The architecture typically consists of an encoder\u2013decoder structure, where encoders extract meaningful representations from input data and decoders generate output sequences based on those representations.\n\nTransformers also use positional encoding to retain information about the order of tokens, since the model itself does not process data sequentially. Combined with multi-head attention and feed-forward neural networks, this design enables scalable training on large datasets.\n\nDue to their flexibility and performance, transformers form the backbone of many modern AI systems, including large language models, machine translation systems, text summarization tools, and increasingly, applications in vision, speech, and multimodal learning.", | |
| "keywords_found": [ | |
| "Multimodal LLMs" | |
| ], | |
| "module_scores": [ | |
| 0.433811753988266, | |
| 0.4582192003726959, | |
| 0.4164702296257019, | |
| 0.2946617603302002, | |
| 0.37847253680229187, | |
| 0.2834630310535431, | |
| 0.2269679754972458, | |
| 0.23173178732395172, | |
| 0.12193305790424347, | |
| 0.15064716339111328, | |
| 0.20222218334674835, | |
| 0.23586700856685638, | |
| 0.2287927120923996, | |
| 0.2737714946269989, | |
| 0.7030925154685974, | |
| 0.5322507619857788, | |
| 0.2212483286857605, | |
| 0.202763170003891 | |
| ], | |
| "strengths": [], | |
| "dominant_topics": [ | |
| "Multimodal LLMs", | |
| "Vision Language Models", | |
| "Pre trained models" | |
| ], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with Multimodal LLMs, Vision Language Models. You successfully reinforced concepts in . Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_32": { | |
| "id": "polyline_32", | |
| "name": "Intro", | |
| "path": [], | |
| "color": "rgba(65, 189, 58, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "Transformers are a class of deep learning models designed to process sequential data efficiently by relying on an attention-based mechanism rather than recurrence or convolution. Introduced in the landmark paper \u201cAttention Is All You Need\u201d, transformers revolutionized natural language processing by enabling models to capture long-range dependencies in data with high parallelism.\n\nAt the core of a transformer is the self-attention mechanism, which allows the model to weigh the importance of different parts of an input sequence when processing each element. This makes transformers highly effective at understanding context, relationships, and structure within data. The architecture typically consists of an encoder\u2013decoder structure, where encoders extract meaningful representations from input data and decoders generate output sequences based on those representations.\n\nTransformers also use positional encoding to retain information about the order of tokens, since the model itself does not process data sequentially. Combined with multi-head attention and feed-forward neural networks, this design enables scalable training on large datasets.\n\nDue to their flexibility and performance, transformers form the backbone of many modern AI systems, including large language models, machine translation systems, text summarization tools, and increasingly, applications in vision, speech, and multimodal learning.", | |
| "keywords_found": [ | |
| "Multimodal LLMs" | |
| ], | |
| "module_scores": [ | |
| 0.433811753988266, | |
| 0.4582192003726959, | |
| 0.4164702296257019, | |
| 0.2946617603302002, | |
| 0.37847253680229187, | |
| 0.2834630310535431, | |
| 0.2269679754972458, | |
| 0.23173178732395172, | |
| 0.12193305790424347, | |
| 0.15064716339111328, | |
| 0.20222218334674835, | |
| 0.23586700856685638, | |
| 0.2287927120923996, | |
| 0.2737714946269989, | |
| 0.7030925154685974, | |
| 0.5322507619857788, | |
| 0.2212483286857605, | |
| 0.202763170003891 | |
| ], | |
| "strengths": [], | |
| "dominant_topics": [ | |
| "Multimodal LLMs", | |
| "Vision Language Models", | |
| "Pre trained models" | |
| ], | |
| "ai_analysis": "To enable advanced AI analysis, set the GEMINI_API_KEY environment variable. Based on your path, you have shown strong engagement with Multimodal LLMs, Vision Language Models. You successfully reinforced concepts in . Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_33": { | |
| "id": "polyline_33", | |
| "name": "Intro", | |
| "path": [ | |
| { | |
| "x": 5, | |
| "y": 16 | |
| } | |
| ], | |
| "color": "rgba(196, 90, 89, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "Transformers are a class of deep learning models designed to process sequential data efficiently by relying on an attention-based mechanism rather than recurrence or convolution. Introduced in the landmark paper \u201cAttention Is All You Need\u201d, transformers revolutionized natural language processing by enabling models to capture long-range dependencies in data with high parallelism.\n\nAt the core of a transformer is the self-attention mechanism, which allows the model to weigh the importance of different parts of an input sequence when processing each element. This makes transformers highly effective at understanding context, relationships, and structure within data. The architecture typically consists of an encoder\u2013decoder structure, where encoders extract meaningful representations from input data and decoders generate output sequences based on those representations.\n\nTransformers also use positional encoding to retain information about the order of tokens, since the model itself does not process data sequentially. Combined with multi-head attention and feed-forward neural networks, this design enables scalable training on large datasets.\n\nDue to their flexibility and performance, transformers form the backbone of many modern AI systems, including large language models, machine translation systems, text summarization tools, and increasingly, applications in vision, speech, and multimodal learning.", | |
| "keywords_found": [ | |
| "Multimodal LLMs" | |
| ], | |
| "module_scores": [ | |
| 0.433811753988266, | |
| 0.4582192003726959, | |
| 0.4164702296257019, | |
| 0.2946617603302002, | |
| 0.37847253680229187, | |
| 0.2834630310535431, | |
| 0.2269679754972458, | |
| 0.23173178732395172, | |
| 0.12193305790424347, | |
| 0.15064716339111328, | |
| 0.20222218334674835, | |
| 0.23586700856685638, | |
| 0.2287927120923996, | |
| 0.2737714946269989, | |
| 0.8030925393104553, | |
| 0.5322507619857788, | |
| 0.2212483286857605, | |
| 0.202763170003891 | |
| ], | |
| "strengths": [ | |
| "Multimodal LLMs" | |
| ], | |
| "dominant_topics": [ | |
| "Multimodal LLMs", | |
| "Vision Language Models", | |
| "Pre trained models" | |
| ], | |
| "ai_analysis": "To enable advanced AI analysis, set the GEMINI_API_KEY environment variable. Based on your path, you have shown strong engagement with Multimodal LLMs, Vision Language Models. You successfully reinforced concepts in Multimodal LLMs. Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_34": { | |
| "id": "polyline_34", | |
| "name": "Pre training LLM", | |
| "path": [ | |
| { | |
| "x": 5, | |
| "y": 17 | |
| } | |
| ], | |
| "color": "rgba(231, 183, 145, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "I have learnt to finetune a pre trained BERT GPT model and i am using these models for sentiment analysis task", | |
| "keywords_found": [ | |
| "Pre trained models" | |
| ], | |
| "module_scores": [ | |
| 0.43252697587013245, | |
| 0.6987414360046387, | |
| 0.2524697184562683, | |
| 0.4963360130786896, | |
| 0.26733019948005676, | |
| 0.28303566575050354, | |
| 0.11330951005220413, | |
| 0.2320728749036789, | |
| 0.07975509762763977, | |
| 0.13868987560272217, | |
| 0.21347038447856903, | |
| 0.13678917288780212, | |
| 0.20242607593536377, | |
| 0.20470771193504333, | |
| 0.19388322532176971, | |
| 0.24398048222064972, | |
| 0.25268879532814026, | |
| 0.3274286389350891 | |
| ], | |
| "strengths": [ | |
| "Agentic AI" | |
| ], | |
| "dominant_topics": [ | |
| "Pre trained models", | |
| "Fine tuning LLM", | |
| "Pre training objectives" | |
| ], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with Pre trained models, Fine tuning LLM. You successfully reinforced concepts in Agentic AI. Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_35": { | |
| "id": "polyline_35", | |
| "name": "heyy", | |
| "path": [ | |
| { | |
| "x": 5, | |
| "y": 17 | |
| } | |
| ], | |
| "color": "rgba(95, 72, 171, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.017389550805091858, | |
| 0.10327564924955368, | |
| 0.09360406547784805, | |
| 0.03690618276596069, | |
| 0.04519070312380791, | |
| 0.038354936987161636, | |
| 0.07477933913469315, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 0.1020158976316452, | |
| 0.07473935186862946, | |
| 0.0972217470407486, | |
| 0.0, | |
| 0.05581950396299362 | |
| ], | |
| "strengths": [ | |
| "Agentic AI" | |
| ], | |
| "dominant_topics": [], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with . You successfully reinforced concepts in Agentic AI. Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_36": { | |
| "id": "polyline_36", | |
| "name": "Agentic AI", | |
| "path": [ | |
| { | |
| "x": 5, | |
| "y": 17 | |
| } | |
| ], | |
| "color": "rgba(139, 171, 145, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "agents independently solve the problem, each can have differnt loss function", | |
| "keywords_found": [ | |
| "Agentic AI" | |
| ], | |
| "module_scores": [ | |
| 0.05024600028991699, | |
| 0.0863848477602005, | |
| 0.0, | |
| 0.1774979680776596, | |
| 0.007330421358346939, | |
| 0.06964121013879776, | |
| 0.08910828828811646, | |
| 0.1227140724658966, | |
| 0.0983947366476059, | |
| 0.02778339385986328, | |
| 0.012763900682330132, | |
| 0.0882045105099678, | |
| 0.2467341423034668, | |
| 0.6995577812194824, | |
| 0.13445383310317993, | |
| 0.07179756462574005, | |
| 0.3390352427959442, | |
| 0.24001526832580566 | |
| ], | |
| "strengths": [ | |
| "Agentic AI" | |
| ], | |
| "dominant_topics": [ | |
| "Agentic AI", | |
| "Policy learning using DQN" | |
| ], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with Agentic AI, Policy learning using DQN. You successfully reinforced concepts in Agentic AI. Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_37": { | |
| "id": "polyline_37", | |
| "name": "Agentic AI, Transformer, CNN", | |
| "path": [ | |
| { | |
| "x": 7, | |
| "y": 16 | |
| }, | |
| { | |
| "x": 5, | |
| "y": 17 | |
| } | |
| ], | |
| "color": "rgba(124, 155, 146, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.73, | |
| "summary": "Agentic AI, AI, Artificial Intelligence, Agentic AI has agents, Agents are good, Agentic AI, Agent", | |
| "keywords_found": [ | |
| "Agentic AI" | |
| ], | |
| "module_scores": [ | |
| 0.19596125185489655, | |
| 0.17878513038158417, | |
| 0.013299948535859585, | |
| 0.09019515663385391, | |
| 0.2148519903421402, | |
| 0.176703080534935, | |
| 0.04141182824969292, | |
| 0.12695123255252838, | |
| 0.14567424356937408, | |
| 0.06627196818590164, | |
| 0.07899114489555359, | |
| 0.17586567997932434, | |
| 0.19476987421512604, | |
| 0.9779932498931885, | |
| 0.14229676127433777, | |
| 0.17270609736442566, | |
| 0.3035743236541748, | |
| 0.3211010694503784 | |
| ], | |
| "strengths": [ | |
| "Quantization", | |
| "Agentic AI" | |
| ], | |
| "dominant_topics": [ | |
| "Agentic AI", | |
| "RLHF", | |
| "Policy learning using DQN" | |
| ], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with Agentic AI, RLHF. You successfully reinforced concepts in Quantization, Agentic AI. Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_38": { | |
| "id": "polyline_38", | |
| "name": "heyy", | |
| "path": [ | |
| { | |
| "x": 7, | |
| "y": 16 | |
| }, | |
| { | |
| "x": 5, | |
| "y": 17 | |
| } | |
| ], | |
| "color": "rgba(198, 232, 160, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.73, | |
| "summary": "heyey", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.03077820874750614, | |
| 0.08723001182079315, | |
| 0.17077267169952393, | |
| 0.0844632163643837, | |
| 0.05345889925956726, | |
| 0.06896784901618958, | |
| 0.01981256529688835, | |
| 0.10914305597543716, | |
| 0.025269674137234688, | |
| 0.0, | |
| 0.0, | |
| 0.12327592819929123, | |
| 0.08018720149993896, | |
| 0.1916126310825348, | |
| 0.07461867481470108, | |
| 0.14787648618221283, | |
| 0.01713082380592823, | |
| 0.15488959848880768 | |
| ], | |
| "strengths": [ | |
| "Quantization", | |
| "Agentic AI" | |
| ], | |
| "dominant_topics": [], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with . You successfully reinforced concepts in Quantization, Agentic AI. Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_39": { | |
| "id": "polyline_39", | |
| "name": "heyu", | |
| "path": [], | |
| "color": "rgba(215, 166, 66, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "heyu", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.025842785835266113, | |
| 0.06289363652467728, | |
| 0.10655583441257477, | |
| 0.059236228466033936, | |
| 0.04497949779033661, | |
| 0.05150831118226051, | |
| 0.08413022756576538, | |
| 0.10349386930465698, | |
| 0.019398359581828117, | |
| 0.023599496111273766, | |
| 0.0, | |
| 0.02620738558471203, | |
| 0.055542171001434326, | |
| 0.08238955587148666, | |
| 0.07436030358076096, | |
| 0.08257163316011429, | |
| 0.05098576098680496, | |
| 0.10993840545415878 | |
| ], | |
| "strengths": [], | |
| "dominant_topics": [], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with . You successfully reinforced concepts in . Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_40": { | |
| "id": "polyline_40", | |
| "name": "he7yey", | |
| "path": [ | |
| { | |
| "x": 4, | |
| "y": 2 | |
| }, | |
| { | |
| "x": 10, | |
| "y": 10 | |
| } | |
| ], | |
| "color": "rgba(68, 203, 241, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.73, | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.017389550805091858, | |
| 0.10327564924955368, | |
| 0.09360406547784805, | |
| 0.03690618276596069, | |
| 0.04519070312380791, | |
| 0.038354936987161636, | |
| 0.17477934062480927, | |
| 0.005562630016356707, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.05210745334625244, | |
| 0.002015892416238785, | |
| 0.07473935186862946, | |
| 0.1972217559814453, | |
| 0.0, | |
| 0.05581950396299362 | |
| ], | |
| "strengths": [ | |
| "Incontext Learning", | |
| "Vision Language Models" | |
| ], | |
| "dominant_topics": [], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with . You successfully reinforced concepts in Incontext Learning, Vision Language Models. Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_41": { | |
| "id": "polyline_41", | |
| "name": "heyy", | |
| "path": [], | |
| "color": "rgba(129, 154, 197, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.7, | |
| "summary": "hyeyyyhu", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.053188253194093704, | |
| 0.0, | |
| 0.04981702193617821, | |
| 0.0, | |
| 0.07837709784507751, | |
| 0.0, | |
| 0.11811202019453049, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0, | |
| 0.0558195635676384, | |
| 0.02478315681219101, | |
| 0.019923537969589233, | |
| 0.0004506018303800374, | |
| 0.046636879444122314 | |
| ], | |
| "strengths": [], | |
| "dominant_topics": [], | |
| "ai_analysis": "Based on your path, you have shown strong engagement with . You successfully reinforced concepts in . Consider exploring advanced topics in new areas next." | |
| }, | |
| "polyline_42": { | |
| "id": "polyline_42", | |
| "name": "heyyu", | |
| "path": [ | |
| { | |
| "x": 10, | |
| "y": 13 | |
| } | |
| ], | |
| "color": "rgba(253, 226, 126, 0.4)", | |
| "isActive": true, | |
| "confidence": 0.71, | |
| "summary": "heyuuu", | |
| "keywords_found": [], | |
| "module_scores": [ | |
| 0.0, | |
| 0.0, | |
| 0.1106286495923996, | |
| 0.04658149555325508, | |
| 0.027793176472187042, | |
| 0.0, | |
| 0.05434393510222435, | |
| 0.037776097655296326, | |
| 0.1, | |
| 0.004652492236346006, | |
| 0.004580324981361628, | |
| 0.01859883777797222, | |
| 0.04390064999461174, | |
| 0.0656537190079689, | |
| 0.0459163598716259, | |
| 0.04920380190014839, | |
| 0.037917736917734146, | |
| 0.09046495705842972 | |
| ], | |
| "strengths": [ | |
| "Prompting methods" | |
| ], | |
| "dominant_topics": [], | |
| "ai_analysis": "AI Insight: Tell student to improve summary about Retrieval Augmented Generation. be critical." | |
| } | |
| }, | |
| "summaries": [ | |
| { | |
| "id": "summary_default_5", | |
| "title": "Test Summary", | |
| "summary": "I learned about BERT and Transformers.", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 2, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Pre training objectives", | |
| "Pre trained models" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 100 | |
| }, | |
| { | |
| "id": "summary_default_6", | |
| "title": "work", | |
| "summary": "I have learned a lot about rag and preprocessing it was pretty good understanding.", | |
| "keywords_found": [ | |
| "Retrieval Augmented Generation" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 4, | |
| "currentLevel": 2, | |
| "strengths": [ | |
| "Prompt based learning", | |
| "Incontext Learning", | |
| "Retrieval Methods", | |
| "Retrieval Augmented Generation" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 200 | |
| }, | |
| { | |
| "id": "summary_default_7", | |
| "title": "Hey", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_8", | |
| "title": "DQN Learning", | |
| "summary": "I studied Policy learning using DQN and reinforcement learning agents.", | |
| "keywords_found": [ | |
| "Agentic AI", | |
| "Policy learning using DQN" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 2, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Pre training objectives", | |
| "Pre trained models" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 100 | |
| }, | |
| { | |
| "id": "summary_default_9", | |
| "title": "heyyy", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_10", | |
| "title": "hii", | |
| "summary": "hiii", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_11", | |
| "title": "hii", | |
| "summary": "nocajco", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_12", | |
| "title": "hii", | |
| "summary": "heyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_13", | |
| "title": "hello", | |
| "summary": "rag and everything was good", | |
| "keywords_found": [ | |
| "Retrieval Augmented Generation" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Retrieval Methods" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_14", | |
| "title": "hello", | |
| "summary": "rag and many other things such as pretrainig objectives and all were good", | |
| "keywords_found": [ | |
| "Pre training objectives", | |
| "Retrieval Augmented Generation" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Pre training objectives" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_15", | |
| "title": "hey ", | |
| "summary": "hey", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_16", | |
| "title": "hey", | |
| "summary": "hey", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_17", | |
| "title": "hey", | |
| "summary": "hey", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_18", | |
| "title": "heyy", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_19", | |
| "title": "heyy", | |
| "summary": "I love rag with my pre trining", | |
| "keywords_found": [ | |
| "Retrieval Augmented Generation" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_20", | |
| "title": "huu", | |
| "summary": "huuu", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_21", | |
| "title": "hello", | |
| "summary": "hello", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_22", | |
| "title": "Introduction", | |
| "summary": "hello", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Tutorial: Introduction to huggingface" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_23", | |
| "title": "heyyy", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_24", | |
| "title": "heyyy", | |
| "summary": "heyyyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_25", | |
| "title": "Introduction to transformers", | |
| "summary": "Transformers are a class of deep learning models designed to process sequential data efficiently by relying on an attention-based mechanism rather than recurrence or convolution. Introduced in the landmark paper \u201cAttention Is All You Need\u201d, transformers revolutionized natural language processing by enabling models to capture long-range dependencies in data with high parallelism.\n\nAt the core of a transformer is the self-attention mechanism, which allows the model to weigh the importance of different parts of an input sequence when processing each element. This makes transformers highly effective at understanding context, relationships, and structure within data. The architecture typically consists of an encoder\u2013decoder structure, where encoders extract meaningful representations from input data and decoders generate output sequences based on those representations.\n\nTransformers also use positional encoding to retain information about the order of tokens, since the model itself does not process data sequentially. Combined with multi-head attention and feed-forward neural networks, this design enables scalable training on large datasets.\n\nDue to their flexibility and performance, transformers form the backbone of many modern AI systems, including large language models, machine translation systems, text summarization tools, and increasingly, applications in vision, speech, and multimodal learning.", | |
| "keywords_found": [ | |
| "Multimodal LLMs" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Tutorial: Introduction to huggingface" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_26", | |
| "title": "heyy", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 5, | |
| "currentLevel": 2, | |
| "strengths": [ | |
| "Tutorial: Introduction to huggingface", | |
| "Fine tuning LLM", | |
| "Instruction tuning", | |
| "Parameter efficient fine tuning", | |
| "Multimodal LLMs" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 250 | |
| }, | |
| { | |
| "id": "summary_default_27", | |
| "title": "heyy", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Tutorial: Introduction to huggingface" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_28", | |
| "title": "introduction to my ", | |
| "summary": "transformers", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Tutorial: Introduction to huggingface" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_29", | |
| "title": "hiii", | |
| "summary": "transformers", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 2, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Tutorial: Introduction to huggingface", | |
| "Fine tuning LLM" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 100 | |
| }, | |
| { | |
| "id": "summary_default_30", | |
| "title": "heyy", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 3, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Instruction tuning", | |
| "Parameter efficient fine tuning", | |
| "Multimodal LLMs" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 150 | |
| }, | |
| { | |
| "id": "summary_default_31", | |
| "title": "Introduction to transformers", | |
| "summary": "Transformers are a class of deep learning models designed to process sequential data efficiently by relying on an attention-based mechanism rather than recurrence or convolution. Introduced in the landmark paper \u201cAttention Is All You Need\u201d, transformers revolutionized natural language processing by enabling models to capture long-range dependencies in data with high parallelism.\n\nAt the core of a transformer is the self-attention mechanism, which allows the model to weigh the importance of different parts of an input sequence when processing each element. This makes transformers highly effective at understanding context, relationships, and structure within data. The architecture typically consists of an encoder\u2013decoder structure, where encoders extract meaningful representations from input data and decoders generate output sequences based on those representations.\n\nTransformers also use positional encoding to retain information about the order of tokens, since the model itself does not process data sequentially. Combined with multi-head attention and feed-forward neural networks, this design enables scalable training on large datasets.\n\nDue to their flexibility and performance, transformers form the backbone of many modern AI systems, including large language models, machine translation systems, text summarization tools, and increasingly, applications in vision, speech, and multimodal learning.", | |
| "keywords_found": [ | |
| "Multimodal LLMs" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_32", | |
| "title": "Intro", | |
| "summary": "Transformers are a class of deep learning models designed to process sequential data efficiently by relying on an attention-based mechanism rather than recurrence or convolution. Introduced in the landmark paper \u201cAttention Is All You Need\u201d, transformers revolutionized natural language processing by enabling models to capture long-range dependencies in data with high parallelism.\n\nAt the core of a transformer is the self-attention mechanism, which allows the model to weigh the importance of different parts of an input sequence when processing each element. This makes transformers highly effective at understanding context, relationships, and structure within data. The architecture typically consists of an encoder\u2013decoder structure, where encoders extract meaningful representations from input data and decoders generate output sequences based on those representations.\n\nTransformers also use positional encoding to retain information about the order of tokens, since the model itself does not process data sequentially. Combined with multi-head attention and feed-forward neural networks, this design enables scalable training on large datasets.\n\nDue to their flexibility and performance, transformers form the backbone of many modern AI systems, including large language models, machine translation systems, text summarization tools, and increasingly, applications in vision, speech, and multimodal learning.", | |
| "keywords_found": [ | |
| "Multimodal LLMs" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_33", | |
| "title": "Intro", | |
| "summary": "Transformers are a class of deep learning models designed to process sequential data efficiently by relying on an attention-based mechanism rather than recurrence or convolution. Introduced in the landmark paper \u201cAttention Is All You Need\u201d, transformers revolutionized natural language processing by enabling models to capture long-range dependencies in data with high parallelism.\n\nAt the core of a transformer is the self-attention mechanism, which allows the model to weigh the importance of different parts of an input sequence when processing each element. This makes transformers highly effective at understanding context, relationships, and structure within data. The architecture typically consists of an encoder\u2013decoder structure, where encoders extract meaningful representations from input data and decoders generate output sequences based on those representations.\n\nTransformers also use positional encoding to retain information about the order of tokens, since the model itself does not process data sequentially. Combined with multi-head attention and feed-forward neural networks, this design enables scalable training on large datasets.\n\nDue to their flexibility and performance, transformers form the backbone of many modern AI systems, including large language models, machine translation systems, text summarization tools, and increasingly, applications in vision, speech, and multimodal learning.", | |
| "keywords_found": [ | |
| "Multimodal LLMs" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Multimodal LLMs" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_34", | |
| "title": "Pre training LLM", | |
| "summary": "I have learnt to finetune a pre trained BERT GPT model and i am using these models for sentiment analysis task", | |
| "keywords_found": [ | |
| "Pre trained models" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Agentic AI" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_35", | |
| "title": "heyy", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Agentic AI" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_36", | |
| "title": "Agentic AI", | |
| "summary": "agents independently solve the problem, each can have differnt loss function", | |
| "keywords_found": [ | |
| "Agentic AI" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Agentic AI" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| }, | |
| { | |
| "id": "summary_default_37", | |
| "title": "Agentic AI, Transformer, CNN", | |
| "summary": "Agentic AI, AI, Artificial Intelligence, Agentic AI has agents, Agents are good, Agentic AI, Agent", | |
| "keywords_found": [ | |
| "Agentic AI" | |
| ], | |
| "totalResources": 18, | |
| "visitedResources": 2, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Quantization", | |
| "Agentic AI" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 100 | |
| }, | |
| { | |
| "id": "summary_default_38", | |
| "title": "heyy", | |
| "summary": "heyey", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 2, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Quantization", | |
| "Agentic AI" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 100 | |
| }, | |
| { | |
| "id": "summary_default_39", | |
| "title": "heyu", | |
| "summary": "heyu", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_40", | |
| "title": "he7yey", | |
| "summary": "heyyy", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 2, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Incontext Learning", | |
| "Vision Language Models" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 100 | |
| }, | |
| { | |
| "id": "summary_default_41", | |
| "title": "heyy", | |
| "summary": "hyeyyyhu", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 0, | |
| "currentLevel": 1, | |
| "strengths": [], | |
| "recommendations": [], | |
| "avgDifficulty": 0, | |
| "totalReward": 0 | |
| }, | |
| { | |
| "id": "summary_default_42", | |
| "title": "heyyu", | |
| "summary": "heyuuu", | |
| "keywords_found": [], | |
| "totalResources": 18, | |
| "visitedResources": 1, | |
| "currentLevel": 1, | |
| "strengths": [ | |
| "Prompting methods" | |
| ], | |
| "recommendations": [], | |
| "avgDifficulty": 2.0, | |
| "totalReward": 50 | |
| } | |
| ] | |
| } |