{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":554025025,"defaultBranch":"main","name":"tpp-pytorch-extension","ownerLogin":"libxsmm","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2022-10-19T06:05:58.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/73218023?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1723616475.0","currentOid":""},"activityList":{"items":[{"before":"79027998d60f8dd5d5e7a5de31edc6bbc9dfe1ba","after":"c6518bbe8ce0c21a36490d37f60e2d7ab23837e8","ref":"refs/heads/gnn_infer","pushedAt":"2024-09-10T11:33:55.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ska278","name":null,"path":"/ska278","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8258117?s=80&v=4"},"commit":{"message":"Bunch of minor changes","shortMessageHtmlLink":"Bunch of minor changes"}},{"before":"e8ef49c2304b3bc6013c7e6e1c5a980a8f44fe10","after":"ec5a5ec8d069538d23a7540c74084262554628b0","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-09-04T11:39:46.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Enable lazy blocking in BlockedLinear","shortMessageHtmlLink":"Enable lazy blocking in BlockedLinear"}},{"before":"6e96f7c613145be1a5291c787ef996d2d59ce382","after":"e8ef49c2304b3bc6013c7e6e1c5a980a8f44fe10","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-09-02T04:01:16.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Update libxsmm commit","shortMessageHtmlLink":"Update libxsmm commit"}},{"before":"406d8cc588e55b3043c35036f2a4aeac9eff615e","after":"6e96f7c613145be1a5291c787ef996d2d59ce382","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-08-28T11:33:14.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Add qint4 and flat gemm support","shortMessageHtmlLink":"Add qint4 and flat gemm support"}},{"before":"dc89b7d3ace41df6a55b88f50282b3273f8eb694","after":"79027998d60f8dd5d5e7a5de31edc6bbc9dfe1ba","ref":"refs/heads/gnn_infer","pushedAt":"2024-08-19T13:39:05.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ska278","name":null,"path":"/ska278","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8258117?s=80&v=4"},"commit":{"message":"Bunch of changes to GNN training and eval scripts","shortMessageHtmlLink":"Bunch of changes to GNN training and eval scripts"}},{"before":"ae7350911e0acd90d4b1903e87c79d5227fa1e5c","after":null,"ref":"refs/heads/narendra","pushedAt":"2024-08-14T06:21:15.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"}},{"before":"2a562b04eea7bfb52e54a9e14b12669623d6eec5","after":"9b1ea033e0f78154c797d0dcea1e36af1b27c745","ref":"refs/heads/main","pushedAt":"2024-08-14T06:21:10.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Removed the fused2 code as it is not needed for now (#25)","shortMessageHtmlLink":"Removed the fused2 code as it is not needed for now (#25)"}},{"before":"84d2b91f95bc23f86f19dd58cbfffdf76b3572f5","after":"dc89b7d3ace41df6a55b88f50282b3273f8eb694","ref":"refs/heads/gnn_infer","pushedAt":"2024-08-10T12:14:14.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ska278","name":null,"path":"/ska278","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8258117?s=80&v=4"},"commit":{"message":"Added profiling code to eval","shortMessageHtmlLink":"Added profiling code to eval"}},{"before":"0af1f92acd827f5f19c32d63d348e06f962d6e38","after":"84d2b91f95bc23f86f19dd58cbfffdf76b3572f5","ref":"refs/heads/gnn_infer","pushedAt":"2024-08-06T14:11:29.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ska278","name":null,"path":"/ska278","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8258117?s=80&v=4"},"commit":{"message":"Fixes for qint8 GEMM in libxsmm. Additional minor bug-fixes in IGBH inference code.","shortMessageHtmlLink":"Fixes for qint8 GEMM in libxsmm. Additional minor bug-fixes in IGBH i…"}},{"before":null,"after":"ae7350911e0acd90d4b1903e87c79d5227fa1e5c","ref":"refs/heads/narendra","pushedAt":"2024-08-06T07:07:20.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"narendrachaudhary51","name":"Narendra Chaudhary","path":"/narendrachaudhary51","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/26152063?s=80&v=4"},"commit":{"message":"Removed the fused2 code as it is not needed for now","shortMessageHtmlLink":"Removed the fused2 code as it is not needed for now"}},{"before":"d577a24aaf8fa790fc5e361fa9789be3e00dd5a1","after":"0af1f92acd827f5f19c32d63d348e06f962d6e38","ref":"refs/heads/gnn_infer","pushedAt":"2024-08-05T13:27:55.000Z","pushType":"push","commitsCount":16,"pusher":{"login":"ska278","name":null,"path":"/ska278","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8258117?s=80&v=4"},"commit":{"message":"Updated int8 inference code","shortMessageHtmlLink":"Updated int8 inference code"}},{"before":"70f5c2201a67d1d29181a794fe167480e86462d5","after":"406d8cc588e55b3043c35036f2a4aeac9eff615e","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-30T10:25:11.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"[LLM] Handle positional arguments correctly in model call wrapper","shortMessageHtmlLink":"[LLM] Handle positional arguments correctly in model call wrapper"}},{"before":"053fe7f714af2ae88d5a3754b4d194cb1d63b996","after":"70f5c2201a67d1d29181a794fe167480e86462d5","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-28T12:07:40.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Allow using power of 2 QBS smaller than QINT8_BLOCK_SIZE","shortMessageHtmlLink":"Allow using power of 2 QBS smaller than QINT8_BLOCK_SIZE"}},{"before":"f91ba445c154b3d87269fdbd94eda27c3b48c53d","after":"053fe7f714af2ae88d5a3754b4d194cb1d63b996","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-27T07:21:38.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"[LLM] Enable running multi-instance when MULTI_INSTANCE set to 1","shortMessageHtmlLink":"[LLM] Enable running multi-instance when MULTI_INSTANCE set to 1"}},{"before":"a0cf090b1d0bf073502c56878b5a35bb3a281772","after":"f91ba445c154b3d87269fdbd94eda27c3b48c53d","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-27T06:29:39.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Enable running multi-instance when MULTI_INSTANCE set to 1","shortMessageHtmlLink":"Enable running multi-instance when MULTI_INSTANCE set to 1"}},{"before":"9eeff407c7f2664b429f6b6e530bbf643b060424","after":"a0cf090b1d0bf073502c56878b5a35bb3a281772","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-26T19:19:22.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Fix null attention mask handling for llama","shortMessageHtmlLink":"Fix null attention mask handling for llama"}},{"before":"6ffbe1a1490427380680ec369b93e1317f781d8d","after":"9eeff407c7f2664b429f6b6e530bbf643b060424","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-25T10:06:25.000Z","pushType":"push","commitsCount":3,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Improve int8 brgemm blocking","shortMessageHtmlLink":"Improve int8 brgemm blocking"}},{"before":"fa9db1381ff5c57f702bd3a355fec52d064c5eff","after":"6ffbe1a1490427380680ec369b93e1317f781d8d","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-23T15:46:57.000Z","pushType":"push","commitsCount":4,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Merge branch 'main' into feature_llm_a8w8","shortMessageHtmlLink":"Merge branch 'main' into feature_llm_a8w8"}},{"before":"68295a6c22a5002ce8e3f2f1aac10c08e0fe8bc7","after":"2a562b04eea7bfb52e54a9e14b12669623d6eec5","ref":"refs/heads/main","pushedAt":"2024-07-23T10:46:42.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Fix for aarch64 build","shortMessageHtmlLink":"Fix for aarch64 build"}},{"before":"2e88f4e0e79a0be9b14ec36f017cb2f89bc18c90","after":"d577a24aaf8fa790fc5e361fa9789be3e00dd5a1","ref":"refs/heads/gnn_infer","pushedAt":"2024-07-23T03:26:09.000Z","pushType":"push","commitsCount":14,"pusher":{"login":"ska278","name":null,"path":"/ska278","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8258117?s=80&v=4"},"commit":{"message":"Fixed int8 quantization utils","shortMessageHtmlLink":"Fixed int8 quantization utils"}},{"before":"0c99bd020fea21731e6b55e1bec346802e30c912","after":"fa9db1381ff5c57f702bd3a355fec52d064c5eff","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-19T13:04:24.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Further refactor quantization code","shortMessageHtmlLink":"Further refactor quantization code"}},{"before":"2c8ab512d148f605bb0b4deacc50f3f35c7fe22b","after":"68295a6c22a5002ce8e3f2f1aac10c08e0fe8bc7","ref":"refs/heads/main","pushedAt":"2024-07-19T05:57:29.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Update setup_conda.sh to use pip instead for conda for intel channel","shortMessageHtmlLink":"Update setup_conda.sh to use pip instead for conda for intel channel"}},{"before":"bb80100089fad640f07da6e9df74f388870c4dc7","after":"2e88f4e0e79a0be9b14ec36f017cb2f89bc18c90","ref":"refs/heads/gnn_infer","pushedAt":"2024-07-18T05:44:48.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ska278","name":null,"path":"/ska278","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8258117?s=80&v=4"},"commit":{"message":"Added GAT Inference code","shortMessageHtmlLink":"Added GAT Inference code"}},{"before":"4201afa5c6dfd6971a3c722c507b5d0e1f944685","after":"0c99bd020fea21731e6b55e1bec346802e30c912","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-17T17:29:44.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Further improve qint8 support","shortMessageHtmlLink":"Further improve qint8 support"}},{"before":"c1e6386b5f8e629ac63aca9296a3088d43f1becf","after":"4201afa5c6dfd6971a3c722c507b5d0e1f944685","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-16T17:52:13.000Z","pushType":"push","commitsCount":3,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Fix quant TPP use and some clean up","shortMessageHtmlLink":"Fix quant TPP use and some clean up"}},{"before":"4a8febeab4c57a73cdd1333c6ff1c6adabb13162","after":"2c8ab512d148f605bb0b4deacc50f3f35c7fe22b","ref":"refs/heads/main","pushedAt":"2024-07-16T17:49:01.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Add support for unequal cores per SNC in run_dist_numa.sh","shortMessageHtmlLink":"Add support for unequal cores per SNC in run_dist_numa.sh"}},{"before":null,"after":"bb80100089fad640f07da6e9df74f388870c4dc7","ref":"refs/heads/gnn_infer","pushedAt":"2024-07-11T13:04:00.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"ska278","name":null,"path":"/ska278","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8258117?s=80&v=4"},"commit":{"message":"Add bf8, hf8 support. Fix downconvert code","shortMessageHtmlLink":"Add bf8, hf8 support. Fix downconvert code"}},{"before":"98a868f4ca7279f29e97ff6d762e8dbb4b1f200d","after":"c1e6386b5f8e629ac63aca9296a3088d43f1becf","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-07-04T06:30:27.000Z","pushType":"push","commitsCount":3,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Remove debug code","shortMessageHtmlLink":"Remove debug code"}},{"before":null,"after":"98a868f4ca7279f29e97ff6d762e8dbb4b1f200d","ref":"refs/heads/feature_llm_a8w8","pushedAt":"2024-06-27T18:03:46.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"Add a8w8 related code","shortMessageHtmlLink":"Add a8w8 related code"}},{"before":"99c84f7c971bd7c253e17d4a55d9562b8d2720bb","after":"4a8febeab4c57a73cdd1333c6ff1c6adabb13162","ref":"refs/heads/main","pushedAt":"2024-06-19T03:57:29.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ddkalamk","name":"Dhiraj D Kalamkar","path":"/ddkalamk","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/8791375?s=80&v=4"},"commit":{"message":"[LLM] Fix warnings in LLM code","shortMessageHtmlLink":"[LLM] Fix warnings in LLM code"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"startCursor":"Y3Vyc29yOnYyOpK7MjAyNC0wOS0xMFQxMTozMzo1NS4wMDAwMDBazwAAAASx-ScS","endCursor":"Y3Vyc29yOnYyOpK7MjAyNC0wNi0xOVQwMzo1NzoyOS4wMDAwMDBazwAAAARpHASa"}},"title":"Activity · libxsmm/tpp-pytorch-extension"}