diff --git "a/test_top_repo_data.csv" "b/test_top_repo_data.csv" --- "a/test_top_repo_data.csv" +++ "b/test_top_repo_data.csv" @@ -1,147 +1,147 @@ Unnamed: 0,id,type,created_at,repo,repo_url,action,title,labels,body,index,text_combine,label,text,binary_label -241762,18474153624.0,IssuesEvent,2021-10-18 04:07:16,numpy/numpy,https://api.github.com/repos/numpy/numpy,closed,DOC: Clarify that numpy average is the mean?,04 - Documentation,"Looking at the documentation for numpy average: -https://github.com/numpy/numpy/blob/b235f9e701e14ed6f6f6dcba885f7986a833743f/numpy/lib/function_base.py#L280-L419 - -I think this is the weighted arithmetic mean, as described at: -https://en.wikipedia.org/wiki/Weighted_arithmetic_mean -Not to be confused with the weighted median: -https://en.wikipedia.org/wiki/Weighted_median - -I suggest clarifying the documentation to clearly say that the weighted mean is being calculated. It would also be nice if the weighted median was included as well.",1.0,"DOC: Clarify that numpy average is the mean? - Looking at the documentation for numpy average: -https://github.com/numpy/numpy/blob/b235f9e701e14ed6f6f6dcba885f7986a833743f/numpy/lib/function_base.py#L280-L419 - -I think this is the weighted arithmetic mean, as described at: -https://en.wikipedia.org/wiki/Weighted_arithmetic_mean -Not to be confused with the weighted median: -https://en.wikipedia.org/wiki/Weighted_median - +241762,18474153624.0,IssuesEvent,2021-10-18 04:07:16,numpy/numpy,https://api.github.com/repos/numpy/numpy,closed,DOC: Clarify that numpy average is the mean?,04 - Documentation,"Looking at the documentation for numpy average: +https://github.com/numpy/numpy/blob/b235f9e701e14ed6f6f6dcba885f7986a833743f/numpy/lib/function_base.py#L280-L419 + +I think this is the weighted arithmetic mean, as described at: +https://en.wikipedia.org/wiki/Weighted_arithmetic_mean +Not to be confused with the weighted median: +https://en.wikipedia.org/wiki/Weighted_median + +I suggest clarifying the documentation to clearly say that the weighted mean is being calculated. It would also be nice if the weighted median was included as well.",1.0,"DOC: Clarify that numpy average is the mean? - Looking at the documentation for numpy average: +https://github.com/numpy/numpy/blob/b235f9e701e14ed6f6f6dcba885f7986a833743f/numpy/lib/function_base.py#L280-L419 + +I think this is the weighted arithmetic mean, as described at: +https://en.wikipedia.org/wiki/Weighted_arithmetic_mean +Not to be confused with the weighted median: +https://en.wikipedia.org/wiki/Weighted_median + I suggest clarifying the documentation to clearly say that the weighted mean is being calculated. It would also be nice if the weighted median was included as well.",0,doc clarify that numpy average is the mean looking at the documentation for numpy average i think this is the weighted arithmetic mean as described at not to be confused with the weighted median i suggest clarifying the documentation to clearly say that the weighted mean is being calculated it would also be nice if the weighted median was included as well ,0 -3299,6261267432.0,IssuesEvent,2017-07-14 23:23:17,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,opened,overflow compability,compatibility rc3.1,"Please answer these questions before submitting your issue. Thanks! - -1. What did you do? - -`select -9223372036854775809;` - - -2. What did you expect to see? - -`9223372036854775809` is overflow math.MaxInt64, but mysql convert it to Decimal type. - -``` -mysql> select -9223372036854775809; -Field 1: `-9223372036854775809` -Catalog: `def` -Database: `` -Table: `` -Org_table: `` -Type: NEWDECIMAL -Collation: binary (63) -Length: 20 -Max_length: 20 -Decimals: 0 -Flags: NOT_NULL BINARY NUM - - -+----------------------+ -| -9223372036854775809 | -+----------------------+ -| -9223372036854775809 | -+----------------------+ -1 row in set (0.00 sec) -``` - -3. What did you see instead? - -but TiDB return 9223372036854775807 caused by int overflow - -``` -mysql> select -9223372036854775809; -Field 1: `-9223372036854775809` -Catalog: `def` -Database: `` -Table: `` -Org_table: `` -Type: LONGLONG -Collation: binary (63) -Length: 0 -Max_length: 19 -Decimals: 31 -Flags: BINARY NUM - - -+----------------------+ -| -9223372036854775809 | -+----------------------+ -| 9223372036854775807 | -+----------------------+ -1 row in set (0.00 sec) -``` - -4. What version of TiDB are you using (`tidb-server -V`)? - -",True,"overflow compability - Please answer these questions before submitting your issue. Thanks! - -1. What did you do? - -`select -9223372036854775809;` - - -2. What did you expect to see? - -`9223372036854775809` is overflow math.MaxInt64, but mysql convert it to Decimal type. - -``` -mysql> select -9223372036854775809; -Field 1: `-9223372036854775809` -Catalog: `def` -Database: `` -Table: `` -Org_table: `` -Type: NEWDECIMAL -Collation: binary (63) -Length: 20 -Max_length: 20 -Decimals: 0 -Flags: NOT_NULL BINARY NUM - - -+----------------------+ -| -9223372036854775809 | -+----------------------+ -| -9223372036854775809 | -+----------------------+ -1 row in set (0.00 sec) -``` - -3. What did you see instead? - -but TiDB return 9223372036854775807 caused by int overflow - -``` -mysql> select -9223372036854775809; -Field 1: `-9223372036854775809` -Catalog: `def` -Database: `` -Table: `` -Org_table: `` -Type: LONGLONG -Collation: binary (63) -Length: 0 -Max_length: 19 -Decimals: 31 -Flags: BINARY NUM - - -+----------------------+ -| -9223372036854775809 | -+----------------------+ -| 9223372036854775807 | -+----------------------+ -1 row in set (0.00 sec) -``` - -4. What version of TiDB are you using (`tidb-server -V`)? - +3299,6261267432.0,IssuesEvent,2017-07-14 23:23:17,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,opened,overflow compability,compatibility rc3.1,"Please answer these questions before submitting your issue. Thanks! + +1. What did you do? + +`select -9223372036854775809;` + + +2. What did you expect to see? + +`9223372036854775809` is overflow math.MaxInt64, but mysql convert it to Decimal type. + +``` +mysql> select -9223372036854775809; +Field 1: `-9223372036854775809` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: NEWDECIMAL +Collation: binary (63) +Length: 20 +Max_length: 20 +Decimals: 0 +Flags: NOT_NULL BINARY NUM + + ++----------------------+ +| -9223372036854775809 | ++----------------------+ +| -9223372036854775809 | ++----------------------+ +1 row in set (0.00 sec) +``` + +3. What did you see instead? + +but TiDB return 9223372036854775807 caused by int overflow + +``` +mysql> select -9223372036854775809; +Field 1: `-9223372036854775809` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: LONGLONG +Collation: binary (63) +Length: 0 +Max_length: 19 +Decimals: 31 +Flags: BINARY NUM + + ++----------------------+ +| -9223372036854775809 | ++----------------------+ +| 9223372036854775807 | ++----------------------+ +1 row in set (0.00 sec) +``` + +4. What version of TiDB are you using (`tidb-server -V`)? + +",True,"overflow compability - Please answer these questions before submitting your issue. Thanks! + +1. What did you do? + +`select -9223372036854775809;` + + +2. What did you expect to see? + +`9223372036854775809` is overflow math.MaxInt64, but mysql convert it to Decimal type. + +``` +mysql> select -9223372036854775809; +Field 1: `-9223372036854775809` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: NEWDECIMAL +Collation: binary (63) +Length: 20 +Max_length: 20 +Decimals: 0 +Flags: NOT_NULL BINARY NUM + + ++----------------------+ +| -9223372036854775809 | ++----------------------+ +| -9223372036854775809 | ++----------------------+ +1 row in set (0.00 sec) +``` + +3. What did you see instead? + +but TiDB return 9223372036854775807 caused by int overflow + +``` +mysql> select -9223372036854775809; +Field 1: `-9223372036854775809` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: LONGLONG +Collation: binary (63) +Length: 0 +Max_length: 19 +Decimals: 31 +Flags: BINARY NUM + + ++----------------------+ +| -9223372036854775809 | ++----------------------+ +| 9223372036854775807 | ++----------------------+ +1 row in set (0.00 sec) +``` + +4. What version of TiDB are you using (`tidb-server -V`)? + ",1,overflow compability please answer these questions before submitting your issue thanks what did you do select what did you expect to see is overflow math but mysql convert it to decimal type mysql select field catalog def database table org table type newdecimal collation binary length max length decimals flags not null binary num row in set sec what did you see instead but tidb return caused by int overflow mysql select field catalog def database table org table type longlong collation binary length max length decimals flags binary num row in set sec what version of tidb are you using tidb server v ,1 207937,15860455507.0,IssuesEvent,2021-04-08 09:10:57,cockroachdb/cockroach,https://api.github.com/repos/cockroachdb/cockroach,opened,roachtest: restore2TB/nodes=10 failed,C-test-failure O-roachtest O-robot branch-release-21.1 release-blocker,"[(roachtest).restore2TB/nodes=10 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2859883&tab=buildLog) on [release-21.1@389cbd4be0e9ce22ca7789cd61802f1f90392c97](https://github.com/cockroachdb/cockroach/commits/389cbd4be0e9ce22ca7789cd61802f1f90392c97): @@ -274,179 +274,179 @@ Related: [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Arestore2TB%2Fnodes%3D10.%2A&sort=title&restgroup=false&display=lastcommented+project) powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
",0,roachtest nodes failed on cluster go restore go cluster go errgroup go output in run cockroach sql home agent work go src github com cockroachdb cockroach bin roachprod run teamcity cockroach sql insecure e context canceled attached stack trace stack trace main cluster rune home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main cluster run home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerrestore home agent work go src github com cockroachdb cockroach pkg cmd roachtest restore go main monitor go home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go golang org x sync errgroup group go home agent work go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go runtime goexit usr local go src runtime asm s wraps output in run cockroach sql wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity cockroach sql insecure e restore csv bank from gs cockroach fixtures workload bank version payload bytes ranges rows seed bank with into db returned stderr stdout wraps secondary error attachment signal killed signal killed error types exec exiterror wraps context canceled error types withstack withstack errutil withprefix main withcommanddetails secondary withsecondaryerror errors errorstring cluster go restore go test runner go monitor failure monitor task failed context canceled attached stack trace stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerrestore home agent work go src github com cockroachdb cockroach pkg cmd roachtest restore go wraps monitor failure wraps attached stack trace stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime goexit usr local go src runtime asm s wraps monitor task failed wraps context canceled error types withstack withstack errutil withprefix withstack withstack errutil withprefix errors errorstring more artifacts related roachtest nodes failed powered by ,0 -3592,6505422239.0,IssuesEvent,2017-08-24 02:51:00,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,[compatibility] SYSDATE() is different from mysql,compatibility,"1. What did you do? -If possible, provide a recipe for reproducing the error. -`SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE();` - -2. What did you expect to see? -``` -mysql> SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE(); -+---------------------+----------+---------------------+----------+---------------------+ -| SYSDATE() | SLEEP(1) | SYSDATE() | SLEEP(2) | SYSDATE() | -+---------------------+----------+---------------------+----------+---------------------+ -| 2017-08-18 22:33:51 | 0 | 2017-08-18 22:33:52 | 0 | 2017-08-18 22:33:54 | -+---------------------+----------+---------------------+----------+---------------------+ -1 row in set (3.01 sec) -``` - - -3. What did you see instead? - -in TiDB SYSDATE() behaves the same as now(), it returns statement begin time but not execute time -``` -MySQL [(none)]> SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE(); -+---------------------+----------+---------------------+----------+---------------------+ -| SYSDATE() | SLEEP(1) | SYSDATE() | SLEEP(2) | SYSDATE() | -+---------------------+----------+---------------------+----------+---------------------+ -| 2017-08-18 22:33:44 | 0 | 2017-08-18 22:33:44 | 0 | 2017-08-18 22:33:44 | -+---------------------+----------+---------------------+----------+---------------------+ -1 row in set (3.00 sec) -``` - - -4. What version of TiDB are you using (`tidb-server -V`)? -Git Commit Hash: 365e4ab491b683a69c2004f30b1ae4f86256e2ad -Git Commit Branch: master -UTC Build Time: 2017-08-16 01:24:27 -",True,"[compatibility] SYSDATE() is different from mysql - 1. What did you do? -If possible, provide a recipe for reproducing the error. -`SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE();` - -2. What did you expect to see? -``` -mysql> SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE(); -+---------------------+----------+---------------------+----------+---------------------+ -| SYSDATE() | SLEEP(1) | SYSDATE() | SLEEP(2) | SYSDATE() | -+---------------------+----------+---------------------+----------+---------------------+ -| 2017-08-18 22:33:51 | 0 | 2017-08-18 22:33:52 | 0 | 2017-08-18 22:33:54 | -+---------------------+----------+---------------------+----------+---------------------+ -1 row in set (3.01 sec) -``` - - -3. What did you see instead? - -in TiDB SYSDATE() behaves the same as now(), it returns statement begin time but not execute time -``` -MySQL [(none)]> SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE(); -+---------------------+----------+---------------------+----------+---------------------+ -| SYSDATE() | SLEEP(1) | SYSDATE() | SLEEP(2) | SYSDATE() | -+---------------------+----------+---------------------+----------+---------------------+ -| 2017-08-18 22:33:44 | 0 | 2017-08-18 22:33:44 | 0 | 2017-08-18 22:33:44 | -+---------------------+----------+---------------------+----------+---------------------+ -1 row in set (3.00 sec) -``` - - -4. What version of TiDB are you using (`tidb-server -V`)? -Git Commit Hash: 365e4ab491b683a69c2004f30b1ae4f86256e2ad -Git Commit Branch: master -UTC Build Time: 2017-08-16 01:24:27 +3592,6505422239.0,IssuesEvent,2017-08-24 02:51:00,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,[compatibility] SYSDATE() is different from mysql,compatibility,"1. What did you do? +If possible, provide a recipe for reproducing the error. +`SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE();` + +2. What did you expect to see? +``` +mysql> SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE(); ++---------------------+----------+---------------------+----------+---------------------+ +| SYSDATE() | SLEEP(1) | SYSDATE() | SLEEP(2) | SYSDATE() | ++---------------------+----------+---------------------+----------+---------------------+ +| 2017-08-18 22:33:51 | 0 | 2017-08-18 22:33:52 | 0 | 2017-08-18 22:33:54 | ++---------------------+----------+---------------------+----------+---------------------+ +1 row in set (3.01 sec) +``` + + +3. What did you see instead? + +in TiDB SYSDATE() behaves the same as now(), it returns statement begin time but not execute time +``` +MySQL [(none)]> SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE(); ++---------------------+----------+---------------------+----------+---------------------+ +| SYSDATE() | SLEEP(1) | SYSDATE() | SLEEP(2) | SYSDATE() | ++---------------------+----------+---------------------+----------+---------------------+ +| 2017-08-18 22:33:44 | 0 | 2017-08-18 22:33:44 | 0 | 2017-08-18 22:33:44 | ++---------------------+----------+---------------------+----------+---------------------+ +1 row in set (3.00 sec) +``` + + +4. What version of TiDB are you using (`tidb-server -V`)? +Git Commit Hash: 365e4ab491b683a69c2004f30b1ae4f86256e2ad +Git Commit Branch: master +UTC Build Time: 2017-08-16 01:24:27 +",True,"[compatibility] SYSDATE() is different from mysql - 1. What did you do? +If possible, provide a recipe for reproducing the error. +`SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE();` + +2. What did you expect to see? +``` +mysql> SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE(); ++---------------------+----------+---------------------+----------+---------------------+ +| SYSDATE() | SLEEP(1) | SYSDATE() | SLEEP(2) | SYSDATE() | ++---------------------+----------+---------------------+----------+---------------------+ +| 2017-08-18 22:33:51 | 0 | 2017-08-18 22:33:52 | 0 | 2017-08-18 22:33:54 | ++---------------------+----------+---------------------+----------+---------------------+ +1 row in set (3.01 sec) +``` + + +3. What did you see instead? + +in TiDB SYSDATE() behaves the same as now(), it returns statement begin time but not execute time +``` +MySQL [(none)]> SELECT SYSDATE(), SLEEP(1), SYSDATE(), SLEEP(2), SYSDATE(); ++---------------------+----------+---------------------+----------+---------------------+ +| SYSDATE() | SLEEP(1) | SYSDATE() | SLEEP(2) | SYSDATE() | ++---------------------+----------+---------------------+----------+---------------------+ +| 2017-08-18 22:33:44 | 0 | 2017-08-18 22:33:44 | 0 | 2017-08-18 22:33:44 | ++---------------------+----------+---------------------+----------+---------------------+ +1 row in set (3.00 sec) +``` + + +4. What version of TiDB are you using (`tidb-server -V`)? +Git Commit Hash: 365e4ab491b683a69c2004f30b1ae4f86256e2ad +Git Commit Branch: master +UTC Build Time: 2017-08-16 01:24:27 ",1, sysdate is different from mysql what did you do if possible provide a recipe for reproducing the error select sysdate sleep sysdate sleep sysdate what did you expect to see mysql select sysdate sleep sysdate sleep sysdate sysdate sleep sysdate sleep sysdate row in set sec what did you see instead in tidb sysdate behaves the same as now it returns statement begin time but not execute time mysql select sysdate sleep sysdate sleep sysdate sysdate sleep sysdate sleep sysdate row in set sec what version of tidb are you using tidb server v git commit hash git commit branch master utc build time ,1 -670895,22709354303.0,IssuesEvent,2022-07-05 17:37:03,brave/brave-browser,https://api.github.com/repos/brave/brave-browser,closed,"[Android] Visiting some websites offers to ""Show simplified view"" for page, even when the option is disabled",priority/P4 QA/Yes repros-on-chrome Chromium/waiting upstream OS/Android," - -## Description -For certain pages and/or under certain conditions, users are asked if they want to view a ""Simplified view"" for the page, even if that option is disabled in `Settings --> Accessibility`. While the conditions required make it a rare occurrence, it is worth mentioning, given that if the option is not selected/enabled, you should not be offered the option at all. - -## Steps to reproduce - -1. Make sure Settings > Accessibility > Simplified view for web pages (Offer to show articles in simplified view, when supported) = disabled -2. Make sure menu > Desktop site = disabled -3. Visit https://ibkr.info/node/3290 3 -## Actual result -Offer to `Show simplified page` appears - -## Expected result -`Show simplified page` offer should not appear. - -## Issue reproduces how often -Easily. - -## Version/Channel Information: - - -- Can you reproduce this issue with the current Play Store version? Yes -- Can you reproduce this issue with the current Play Store Beta version? Yes -- Can you reproduce this issue with the current Play Store Nightly version? Yes - -## Brave version -v1.21.76 - -### Website problems only -- Does the issue resolve itself when disabling Brave Shields? N/A -- Does the issue resolve itself when disabling Brave Rewards? N/A -- Is the issue reproducible on the latest version of Chrome? N/A - -### Additional information - -User reports: -https://community.brave.com/t/x-buttons-not-functional-design-and-features-how-to-make-brave-even-more-revolutionary/171930/5 - -Additional steps to reproduce on a different website (taken from user directly): - -> 1. set Settings > Accessibility > Simplified view for webpages = disabled -> 2. close all Private tabs (if you have any) -> 5. open new Private tab -> 6. (this has the same effect as clearing site data for this purpose) -> 8. make sure: menu > Desktop site = disabled -> 10. visit https://en.wikipedia.org/wiki/Auction_theory (it’ll redirect to mobile version) -> 14. Scroll to bottom, tap “Desktop” hyperlink -> 16. In the first paragraph of the article, tap “market failure” hyperlink -> 18. You should see the “Show simplified view” bar when the “market failure” page loads (if not, try going through to a couple more articles). The bar isn’t supposed to be presented to the user since you disabled it in step 1 above. Therefore, it’s a bug. :slight_smile: Now if you tap Back, the bar will appear on previous article also.",1.0,"[Android] Visiting some websites offers to ""Show simplified view"" for page, even when the option is disabled - - -## Description -For certain pages and/or under certain conditions, users are asked if they want to view a ""Simplified view"" for the page, even if that option is disabled in `Settings --> Accessibility`. While the conditions required make it a rare occurrence, it is worth mentioning, given that if the option is not selected/enabled, you should not be offered the option at all. - -## Steps to reproduce - -1. Make sure Settings > Accessibility > Simplified view for web pages (Offer to show articles in simplified view, when supported) = disabled -2. Make sure menu > Desktop site = disabled -3. Visit https://ibkr.info/node/3290 3 -## Actual result -Offer to `Show simplified page` appears - -## Expected result -`Show simplified page` offer should not appear. - -## Issue reproduces how often -Easily. - -## Version/Channel Information: - - -- Can you reproduce this issue with the current Play Store version? Yes -- Can you reproduce this issue with the current Play Store Beta version? Yes -- Can you reproduce this issue with the current Play Store Nightly version? Yes - -## Brave version -v1.21.76 - -### Website problems only -- Does the issue resolve itself when disabling Brave Shields? N/A -- Does the issue resolve itself when disabling Brave Rewards? N/A -- Is the issue reproducible on the latest version of Chrome? N/A - -### Additional information - -User reports: -https://community.brave.com/t/x-buttons-not-functional-design-and-features-how-to-make-brave-even-more-revolutionary/171930/5 - -Additional steps to reproduce on a different website (taken from user directly): - -> 1. set Settings > Accessibility > Simplified view for webpages = disabled -> 2. close all Private tabs (if you have any) -> 5. open new Private tab -> 6. (this has the same effect as clearing site data for this purpose) -> 8. make sure: menu > Desktop site = disabled -> 10. visit https://en.wikipedia.org/wiki/Auction_theory (it’ll redirect to mobile version) -> 14. Scroll to bottom, tap “Desktop” hyperlink -> 16. In the first paragraph of the article, tap “market failure” hyperlink +670895,22709354303.0,IssuesEvent,2022-07-05 17:37:03,brave/brave-browser,https://api.github.com/repos/brave/brave-browser,closed,"[Android] Visiting some websites offers to ""Show simplified view"" for page, even when the option is disabled",priority/P4 QA/Yes repros-on-chrome Chromium/waiting upstream OS/Android," + +## Description +For certain pages and/or under certain conditions, users are asked if they want to view a ""Simplified view"" for the page, even if that option is disabled in `Settings --> Accessibility`. While the conditions required make it a rare occurrence, it is worth mentioning, given that if the option is not selected/enabled, you should not be offered the option at all. + +## Steps to reproduce + +1. Make sure Settings > Accessibility > Simplified view for web pages (Offer to show articles in simplified view, when supported) = disabled +2. Make sure menu > Desktop site = disabled +3. Visit https://ibkr.info/node/3290 3 +## Actual result +Offer to `Show simplified page` appears + +## Expected result +`Show simplified page` offer should not appear. + +## Issue reproduces how often +Easily. + +## Version/Channel Information: + + +- Can you reproduce this issue with the current Play Store version? Yes +- Can you reproduce this issue with the current Play Store Beta version? Yes +- Can you reproduce this issue with the current Play Store Nightly version? Yes + +## Brave version +v1.21.76 + +### Website problems only +- Does the issue resolve itself when disabling Brave Shields? N/A +- Does the issue resolve itself when disabling Brave Rewards? N/A +- Is the issue reproducible on the latest version of Chrome? N/A + +### Additional information + +User reports: +https://community.brave.com/t/x-buttons-not-functional-design-and-features-how-to-make-brave-even-more-revolutionary/171930/5 + +Additional steps to reproduce on a different website (taken from user directly): + +> 1. set Settings > Accessibility > Simplified view for webpages = disabled +> 2. close all Private tabs (if you have any) +> 5. open new Private tab +> 6. (this has the same effect as clearing site data for this purpose) +> 8. make sure: menu > Desktop site = disabled +> 10. visit https://en.wikipedia.org/wiki/Auction_theory (it’ll redirect to mobile version) +> 14. Scroll to bottom, tap “Desktop” hyperlink +> 16. In the first paragraph of the article, tap “market failure” hyperlink +> 18. You should see the “Show simplified view” bar when the “market failure” page loads (if not, try going through to a couple more articles). The bar isn’t supposed to be presented to the user since you disabled it in step 1 above. Therefore, it’s a bug. :slight_smile: Now if you tap Back, the bar will appear on previous article also.",1.0,"[Android] Visiting some websites offers to ""Show simplified view"" for page, even when the option is disabled - + +## Description +For certain pages and/or under certain conditions, users are asked if they want to view a ""Simplified view"" for the page, even if that option is disabled in `Settings --> Accessibility`. While the conditions required make it a rare occurrence, it is worth mentioning, given that if the option is not selected/enabled, you should not be offered the option at all. + +## Steps to reproduce + +1. Make sure Settings > Accessibility > Simplified view for web pages (Offer to show articles in simplified view, when supported) = disabled +2. Make sure menu > Desktop site = disabled +3. Visit https://ibkr.info/node/3290 3 +## Actual result +Offer to `Show simplified page` appears + +## Expected result +`Show simplified page` offer should not appear. + +## Issue reproduces how often +Easily. + +## Version/Channel Information: + + +- Can you reproduce this issue with the current Play Store version? Yes +- Can you reproduce this issue with the current Play Store Beta version? Yes +- Can you reproduce this issue with the current Play Store Nightly version? Yes + +## Brave version +v1.21.76 + +### Website problems only +- Does the issue resolve itself when disabling Brave Shields? N/A +- Does the issue resolve itself when disabling Brave Rewards? N/A +- Is the issue reproducible on the latest version of Chrome? N/A + +### Additional information + +User reports: +https://community.brave.com/t/x-buttons-not-functional-design-and-features-how-to-make-brave-even-more-revolutionary/171930/5 + +Additional steps to reproduce on a different website (taken from user directly): + +> 1. set Settings > Accessibility > Simplified view for webpages = disabled +> 2. close all Private tabs (if you have any) +> 5. open new Private tab +> 6. (this has the same effect as clearing site data for this purpose) +> 8. make sure: menu > Desktop site = disabled +> 10. visit https://en.wikipedia.org/wiki/Auction_theory (it’ll redirect to mobile version) +> 14. Scroll to bottom, tap “Desktop” hyperlink +> 16. In the first paragraph of the article, tap “market failure” hyperlink > 18. You should see the “Show simplified view” bar when the “market failure” page loads (if not, try going through to a couple more articles). The bar isn’t supposed to be presented to the user since you disabled it in step 1 above. Therefore, it’s a bug. :slight_smile: Now if you tap Back, the bar will appear on previous article also.",0, visiting some websites offers to show simplified view for page even when the option is disabled have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description for certain pages and or under certain conditions users are asked if they want to view a simplified view for the page even if that option is disabled in settings accessibility while the conditions required make it a rare occurrence it is worth mentioning given that if the option is not selected enabled you should not be offered the option at all steps to reproduce make sure settings accessibility simplified view for web pages offer to show articles in simplified view when supported disabled make sure menu desktop site disabled visit actual result offer to show simplified page appears expected result show simplified page offer should not appear issue reproduces how often easily version channel information can you reproduce this issue with the current play store version yes can you reproduce this issue with the current play store beta version yes can you reproduce this issue with the current play store nightly version yes brave version website problems only does the issue resolve itself when disabling brave shields n a does the issue resolve itself when disabling brave rewards n a is the issue reproducible on the latest version of chrome n a additional information user reports additional steps to reproduce on a different website taken from user directly set settings accessibility simplified view for webpages disabled close all private tabs if you have any open new private tab this has the same effect as clearing site data for this purpose make sure menu desktop site disabled visit it’ll redirect to mobile version scroll to bottom tap “desktop” hyperlink in the first paragraph of the article tap “market failure” hyperlink you should see the “show simplified view” bar when the “market failure” page loads if not try going through to a couple more articles the bar isn’t supposed to be presented to the user since you disabled it in step above therefore it’s a bug slight smile now if you tap back the bar will appear on previous article also ,0 346139,30869223545.0,IssuesEvent,2023-08-03 10:07:49,ita-social-projects/Space2Study-Client-mvp,https://api.github.com/repos/ita-social-projects/Space2Study-Client-mvp,opened,"(SP: 1) Write unit test for ""GeneralInfoStep"" container",FrontEnd part Unit test,"### Container unit test Unit test for ""**AddPhotoStep**"" container @@ -467,180 +467,180 @@ Scenaries descriptions: Current coverage: ",0, sp write unit test for generalinfostep container container unit test unit test for addphotostep container scenaries descriptions container should rendered check if the buttons passed in props is in the document current coverage ,0 -3849,6705903836.0,IssuesEvent,2017-10-12 03:25:33,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,opened,group by with alias variable not consistent with MySQL.,compatibility,"test.sql: - -``` -DROP TABLE IF EXISTS t1; -SET sql_mode = 'NO_ENGINE_SUBSTITUTION'; -CREATE TABLE t1 ( -cont_nr int(11) NOT NULL auto_increment, -ver_nr int(11) NOT NULL default '0', -aufnr int(11) NOT NULL default '0', -username varchar(50) NOT NULL default '', -hdl_nr int(11) NOT NULL default '0', -eintrag date NOT NULL default '0000-00-00', -st_klasse varchar(40) NOT NULL default '', -st_wert varchar(40) NOT NULL default '', -st_zusatz varchar(40) NOT NULL default '', -st_bemerkung varchar(255) NOT NULL default '', -kunden_art varchar(40) NOT NULL default '', -mcbs_knr int(11) default NULL, -mcbs_aufnr int(11) NOT NULL default '0', -schufa_status char(1) default '?', -bemerkung text, -wirknetz text, -wf_igz int(11) NOT NULL default '0', -tarifcode varchar(80) default NULL, -recycle char(1) default NULL, -sim varchar(30) default NULL, -mcbs_tpl varchar(30) default NULL, -emp_nr int(11) NOT NULL default '0', -laufzeit int(11) default NULL, -hdl_name varchar(30) default NULL, -prov_hdl_nr int(11) NOT NULL default '0', -auto_wirknetz varchar(50) default NULL, -auto_billing varchar(50) default NULL, -touch timestamp NOT NULL, -kategorie varchar(50) default NULL, -kundentyp varchar(20) NOT NULL default '', -sammel_rech_msisdn varchar(30) NOT NULL default '', -p_nr varchar(9) NOT NULL default '', -suffix char(3) NOT NULL default '', -PRIMARY KEY (cont_nr), -KEY idx_aufnr(aufnr), -KEY idx_hdl_nr(hdl_nr), -KEY idx_st_klasse(st_klasse), -KEY ver_nr(ver_nr), -KEY eintrag_idx(eintrag), -KEY emp_nr_idx(emp_nr), -KEY wf_igz(wf_igz), -KEY touch(touch), -KEY hdl_tag(eintrag,hdl_nr), -KEY prov_hdl_nr(prov_hdl_nr), -KEY mcbs_aufnr(mcbs_aufnr), -KEY kundentyp(kundentyp), -KEY p_nr(p_nr,suffix) -); -INSERT INTO t1 VALUES (3359356,405,3359356,'Mustermann Musterfrau',52500,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1485525,2122316,'+','','N',1909160,'MobilComSuper92000D2',NULL,NULL,'MS9ND2',3,24,'MobilCom Shop Koeln',52500,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359357,468,3359357,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1503580,2139699,'+','','P',1909171,'MobilComSuper9D1T10SFreisprech(Akquise)',NULL,NULL,'MS9NS1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359358,407,3359358,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1501358,2137473,'N','','N',1909159,'MobilComSuper92000D2',NULL,NULL,'MS9ND2',325,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359359,468,3359359,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1507831,2143894,'+','','P',1909162,'MobilComSuper9D1T10SFreisprech(Akquise)',NULL,NULL,'MS9NS1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359360,0,0,'Mustermann Musterfrau',29674907,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1900169997,2414578,'+',NULL,'N',1909148,'',NULL,NULL,'RV99066_2',20,NULL,'POS',29674907,NULL,NULL,20010202105916,'Mobilfunk','','','97317481','007'); -INSERT INTO t1 VALUES (3359361,406,3359361,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag storniert','','(7001-84):Storno, Kd. m�chte nicht mehr','privat',NULL,0,'+','','P',1909150,'MobilComSuper92000D1(Akquise)',NULL,NULL,'MS9ND1',325,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359362,406,3359362,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1509984,2145874,'+','','P',1909154,'MobilComSuper92000D1(Akquise)',NULL,NULL,'MS9ND1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -SELECT ELT(FIELD(kundentyp,'PP','PPA','PG','PGA','FK','FKA','FP','FPA','K','KA','V','VA',''), 'Privat (Private Nutzung)','Privat (Private Nutzung) Sitz im Ausland','Privat (geschaeftliche Nutzung)','Privat (geschaeftliche Nutzung) Sitz im Ausland','Firma (Kapitalgesellschaft)','Firma (Kapitalgesellschaft) Sitz im Ausland','Firma (Personengesellschaft)','Firma (Personengesellschaft) Sitz im Ausland','oeff. rechtl. Koerperschaft','oeff. rechtl. Koerperschaft Sitz im Ausland','Eingetragener Verein','Eingetragener Verein Sitz im Ausland','Typ unbekannt') AS Kundentyp ,kategorie FROM t1 WHERE hdl_nr < 2000000 AND kategorie IN ('Prepaid','Mobilfunk') AND st_klasse = 'Workflow' GROUP BY kundentyp ORDER BY kategorie; -``` - -MySQL got: - -``` -+--------------------------+-----------+ -| Kundentyp | kategorie | -+--------------------------+-----------+ -| Privat (Private Nutzung) | Mobilfunk | -+--------------------------+-----------+ -1 row in set, 1 warning (0.00 sec) - -mysql> show warnings; -+---------+------+----------------------------------------------------+ -| Level | Code | Message | -+---------+------+----------------------------------------------------+ -| Warning | 1052 | Column 'kundentyp' in group statement is ambiguous | -+---------+------+----------------------------------------------------+ -1 row in set (0.00 sec) -``` - -but TiDB got empty set; - - - -",True,"group by with alias variable not consistent with MySQL. - test.sql: - -``` -DROP TABLE IF EXISTS t1; -SET sql_mode = 'NO_ENGINE_SUBSTITUTION'; -CREATE TABLE t1 ( -cont_nr int(11) NOT NULL auto_increment, -ver_nr int(11) NOT NULL default '0', -aufnr int(11) NOT NULL default '0', -username varchar(50) NOT NULL default '', -hdl_nr int(11) NOT NULL default '0', -eintrag date NOT NULL default '0000-00-00', -st_klasse varchar(40) NOT NULL default '', -st_wert varchar(40) NOT NULL default '', -st_zusatz varchar(40) NOT NULL default '', -st_bemerkung varchar(255) NOT NULL default '', -kunden_art varchar(40) NOT NULL default '', -mcbs_knr int(11) default NULL, -mcbs_aufnr int(11) NOT NULL default '0', -schufa_status char(1) default '?', -bemerkung text, -wirknetz text, -wf_igz int(11) NOT NULL default '0', -tarifcode varchar(80) default NULL, -recycle char(1) default NULL, -sim varchar(30) default NULL, -mcbs_tpl varchar(30) default NULL, -emp_nr int(11) NOT NULL default '0', -laufzeit int(11) default NULL, -hdl_name varchar(30) default NULL, -prov_hdl_nr int(11) NOT NULL default '0', -auto_wirknetz varchar(50) default NULL, -auto_billing varchar(50) default NULL, -touch timestamp NOT NULL, -kategorie varchar(50) default NULL, -kundentyp varchar(20) NOT NULL default '', -sammel_rech_msisdn varchar(30) NOT NULL default '', -p_nr varchar(9) NOT NULL default '', -suffix char(3) NOT NULL default '', -PRIMARY KEY (cont_nr), -KEY idx_aufnr(aufnr), -KEY idx_hdl_nr(hdl_nr), -KEY idx_st_klasse(st_klasse), -KEY ver_nr(ver_nr), -KEY eintrag_idx(eintrag), -KEY emp_nr_idx(emp_nr), -KEY wf_igz(wf_igz), -KEY touch(touch), -KEY hdl_tag(eintrag,hdl_nr), -KEY prov_hdl_nr(prov_hdl_nr), -KEY mcbs_aufnr(mcbs_aufnr), -KEY kundentyp(kundentyp), -KEY p_nr(p_nr,suffix) -); -INSERT INTO t1 VALUES (3359356,405,3359356,'Mustermann Musterfrau',52500,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1485525,2122316,'+','','N',1909160,'MobilComSuper92000D2',NULL,NULL,'MS9ND2',3,24,'MobilCom Shop Koeln',52500,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359357,468,3359357,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1503580,2139699,'+','','P',1909171,'MobilComSuper9D1T10SFreisprech(Akquise)',NULL,NULL,'MS9NS1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359358,407,3359358,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1501358,2137473,'N','','N',1909159,'MobilComSuper92000D2',NULL,NULL,'MS9ND2',325,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359359,468,3359359,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1507831,2143894,'+','','P',1909162,'MobilComSuper9D1T10SFreisprech(Akquise)',NULL,NULL,'MS9NS1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359360,0,0,'Mustermann Musterfrau',29674907,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1900169997,2414578,'+',NULL,'N',1909148,'',NULL,NULL,'RV99066_2',20,NULL,'POS',29674907,NULL,NULL,20010202105916,'Mobilfunk','','','97317481','007'); -INSERT INTO t1 VALUES (3359361,406,3359361,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag storniert','','(7001-84):Storno, Kd. m�chte nicht mehr','privat',NULL,0,'+','','P',1909150,'MobilComSuper92000D1(Akquise)',NULL,NULL,'MS9ND1',325,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -INSERT INTO t1 VALUES (3359362,406,3359362,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1509984,2145874,'+','','P',1909154,'MobilComSuper92000D1(Akquise)',NULL,NULL,'MS9ND1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); -SELECT ELT(FIELD(kundentyp,'PP','PPA','PG','PGA','FK','FKA','FP','FPA','K','KA','V','VA',''), 'Privat (Private Nutzung)','Privat (Private Nutzung) Sitz im Ausland','Privat (geschaeftliche Nutzung)','Privat (geschaeftliche Nutzung) Sitz im Ausland','Firma (Kapitalgesellschaft)','Firma (Kapitalgesellschaft) Sitz im Ausland','Firma (Personengesellschaft)','Firma (Personengesellschaft) Sitz im Ausland','oeff. rechtl. Koerperschaft','oeff. rechtl. Koerperschaft Sitz im Ausland','Eingetragener Verein','Eingetragener Verein Sitz im Ausland','Typ unbekannt') AS Kundentyp ,kategorie FROM t1 WHERE hdl_nr < 2000000 AND kategorie IN ('Prepaid','Mobilfunk') AND st_klasse = 'Workflow' GROUP BY kundentyp ORDER BY kategorie; -``` - -MySQL got: - -``` -+--------------------------+-----------+ -| Kundentyp | kategorie | -+--------------------------+-----------+ -| Privat (Private Nutzung) | Mobilfunk | -+--------------------------+-----------+ -1 row in set, 1 warning (0.00 sec) - -mysql> show warnings; -+---------+------+----------------------------------------------------+ -| Level | Code | Message | -+---------+------+----------------------------------------------------+ -| Warning | 1052 | Column 'kundentyp' in group statement is ambiguous | -+---------+------+----------------------------------------------------+ -1 row in set (0.00 sec) -``` - -but TiDB got empty set; - - - +3849,6705903836.0,IssuesEvent,2017-10-12 03:25:33,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,opened,group by with alias variable not consistent with MySQL.,compatibility,"test.sql: + +``` +DROP TABLE IF EXISTS t1; +SET sql_mode = 'NO_ENGINE_SUBSTITUTION'; +CREATE TABLE t1 ( +cont_nr int(11) NOT NULL auto_increment, +ver_nr int(11) NOT NULL default '0', +aufnr int(11) NOT NULL default '0', +username varchar(50) NOT NULL default '', +hdl_nr int(11) NOT NULL default '0', +eintrag date NOT NULL default '0000-00-00', +st_klasse varchar(40) NOT NULL default '', +st_wert varchar(40) NOT NULL default '', +st_zusatz varchar(40) NOT NULL default '', +st_bemerkung varchar(255) NOT NULL default '', +kunden_art varchar(40) NOT NULL default '', +mcbs_knr int(11) default NULL, +mcbs_aufnr int(11) NOT NULL default '0', +schufa_status char(1) default '?', +bemerkung text, +wirknetz text, +wf_igz int(11) NOT NULL default '0', +tarifcode varchar(80) default NULL, +recycle char(1) default NULL, +sim varchar(30) default NULL, +mcbs_tpl varchar(30) default NULL, +emp_nr int(11) NOT NULL default '0', +laufzeit int(11) default NULL, +hdl_name varchar(30) default NULL, +prov_hdl_nr int(11) NOT NULL default '0', +auto_wirknetz varchar(50) default NULL, +auto_billing varchar(50) default NULL, +touch timestamp NOT NULL, +kategorie varchar(50) default NULL, +kundentyp varchar(20) NOT NULL default '', +sammel_rech_msisdn varchar(30) NOT NULL default '', +p_nr varchar(9) NOT NULL default '', +suffix char(3) NOT NULL default '', +PRIMARY KEY (cont_nr), +KEY idx_aufnr(aufnr), +KEY idx_hdl_nr(hdl_nr), +KEY idx_st_klasse(st_klasse), +KEY ver_nr(ver_nr), +KEY eintrag_idx(eintrag), +KEY emp_nr_idx(emp_nr), +KEY wf_igz(wf_igz), +KEY touch(touch), +KEY hdl_tag(eintrag,hdl_nr), +KEY prov_hdl_nr(prov_hdl_nr), +KEY mcbs_aufnr(mcbs_aufnr), +KEY kundentyp(kundentyp), +KEY p_nr(p_nr,suffix) +); +INSERT INTO t1 VALUES (3359356,405,3359356,'Mustermann Musterfrau',52500,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1485525,2122316,'+','','N',1909160,'MobilComSuper92000D2',NULL,NULL,'MS9ND2',3,24,'MobilCom Shop Koeln',52500,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359357,468,3359357,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1503580,2139699,'+','','P',1909171,'MobilComSuper9D1T10SFreisprech(Akquise)',NULL,NULL,'MS9NS1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359358,407,3359358,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1501358,2137473,'N','','N',1909159,'MobilComSuper92000D2',NULL,NULL,'MS9ND2',325,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359359,468,3359359,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1507831,2143894,'+','','P',1909162,'MobilComSuper9D1T10SFreisprech(Akquise)',NULL,NULL,'MS9NS1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359360,0,0,'Mustermann Musterfrau',29674907,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1900169997,2414578,'+',NULL,'N',1909148,'',NULL,NULL,'RV99066_2',20,NULL,'POS',29674907,NULL,NULL,20010202105916,'Mobilfunk','','','97317481','007'); +INSERT INTO t1 VALUES (3359361,406,3359361,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag storniert','','(7001-84):Storno, Kd. m�chte nicht mehr','privat',NULL,0,'+','','P',1909150,'MobilComSuper92000D1(Akquise)',NULL,NULL,'MS9ND1',325,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359362,406,3359362,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1509984,2145874,'+','','P',1909154,'MobilComSuper92000D1(Akquise)',NULL,NULL,'MS9ND1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +SELECT ELT(FIELD(kundentyp,'PP','PPA','PG','PGA','FK','FKA','FP','FPA','K','KA','V','VA',''), 'Privat (Private Nutzung)','Privat (Private Nutzung) Sitz im Ausland','Privat (geschaeftliche Nutzung)','Privat (geschaeftliche Nutzung) Sitz im Ausland','Firma (Kapitalgesellschaft)','Firma (Kapitalgesellschaft) Sitz im Ausland','Firma (Personengesellschaft)','Firma (Personengesellschaft) Sitz im Ausland','oeff. rechtl. Koerperschaft','oeff. rechtl. Koerperschaft Sitz im Ausland','Eingetragener Verein','Eingetragener Verein Sitz im Ausland','Typ unbekannt') AS Kundentyp ,kategorie FROM t1 WHERE hdl_nr < 2000000 AND kategorie IN ('Prepaid','Mobilfunk') AND st_klasse = 'Workflow' GROUP BY kundentyp ORDER BY kategorie; +``` + +MySQL got: + +``` ++--------------------------+-----------+ +| Kundentyp | kategorie | ++--------------------------+-----------+ +| Privat (Private Nutzung) | Mobilfunk | ++--------------------------+-----------+ +1 row in set, 1 warning (0.00 sec) + +mysql> show warnings; ++---------+------+----------------------------------------------------+ +| Level | Code | Message | ++---------+------+----------------------------------------------------+ +| Warning | 1052 | Column 'kundentyp' in group statement is ambiguous | ++---------+------+----------------------------------------------------+ +1 row in set (0.00 sec) +``` + +but TiDB got empty set; + + + +",True,"group by with alias variable not consistent with MySQL. - test.sql: + +``` +DROP TABLE IF EXISTS t1; +SET sql_mode = 'NO_ENGINE_SUBSTITUTION'; +CREATE TABLE t1 ( +cont_nr int(11) NOT NULL auto_increment, +ver_nr int(11) NOT NULL default '0', +aufnr int(11) NOT NULL default '0', +username varchar(50) NOT NULL default '', +hdl_nr int(11) NOT NULL default '0', +eintrag date NOT NULL default '0000-00-00', +st_klasse varchar(40) NOT NULL default '', +st_wert varchar(40) NOT NULL default '', +st_zusatz varchar(40) NOT NULL default '', +st_bemerkung varchar(255) NOT NULL default '', +kunden_art varchar(40) NOT NULL default '', +mcbs_knr int(11) default NULL, +mcbs_aufnr int(11) NOT NULL default '0', +schufa_status char(1) default '?', +bemerkung text, +wirknetz text, +wf_igz int(11) NOT NULL default '0', +tarifcode varchar(80) default NULL, +recycle char(1) default NULL, +sim varchar(30) default NULL, +mcbs_tpl varchar(30) default NULL, +emp_nr int(11) NOT NULL default '0', +laufzeit int(11) default NULL, +hdl_name varchar(30) default NULL, +prov_hdl_nr int(11) NOT NULL default '0', +auto_wirknetz varchar(50) default NULL, +auto_billing varchar(50) default NULL, +touch timestamp NOT NULL, +kategorie varchar(50) default NULL, +kundentyp varchar(20) NOT NULL default '', +sammel_rech_msisdn varchar(30) NOT NULL default '', +p_nr varchar(9) NOT NULL default '', +suffix char(3) NOT NULL default '', +PRIMARY KEY (cont_nr), +KEY idx_aufnr(aufnr), +KEY idx_hdl_nr(hdl_nr), +KEY idx_st_klasse(st_klasse), +KEY ver_nr(ver_nr), +KEY eintrag_idx(eintrag), +KEY emp_nr_idx(emp_nr), +KEY wf_igz(wf_igz), +KEY touch(touch), +KEY hdl_tag(eintrag,hdl_nr), +KEY prov_hdl_nr(prov_hdl_nr), +KEY mcbs_aufnr(mcbs_aufnr), +KEY kundentyp(kundentyp), +KEY p_nr(p_nr,suffix) +); +INSERT INTO t1 VALUES (3359356,405,3359356,'Mustermann Musterfrau',52500,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1485525,2122316,'+','','N',1909160,'MobilComSuper92000D2',NULL,NULL,'MS9ND2',3,24,'MobilCom Shop Koeln',52500,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359357,468,3359357,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1503580,2139699,'+','','P',1909171,'MobilComSuper9D1T10SFreisprech(Akquise)',NULL,NULL,'MS9NS1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359358,407,3359358,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1501358,2137473,'N','','N',1909159,'MobilComSuper92000D2',NULL,NULL,'MS9ND2',325,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359359,468,3359359,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1507831,2143894,'+','','P',1909162,'MobilComSuper9D1T10SFreisprech(Akquise)',NULL,NULL,'MS9NS1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359360,0,0,'Mustermann Musterfrau',29674907,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1900169997,2414578,'+',NULL,'N',1909148,'',NULL,NULL,'RV99066_2',20,NULL,'POS',29674907,NULL,NULL,20010202105916,'Mobilfunk','','','97317481','007'); +INSERT INTO t1 VALUES (3359361,406,3359361,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag storniert','','(7001-84):Storno, Kd. m�chte nicht mehr','privat',NULL,0,'+','','P',1909150,'MobilComSuper92000D1(Akquise)',NULL,NULL,'MS9ND1',325,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +INSERT INTO t1 VALUES (3359362,406,3359362,'Mustermann Musterfrau',7001,'2000-05-20','workflow','Auftrag erledigt','Originalvertrag eingegangen und gepr�ft','','privat',1509984,2145874,'+','','P',1909154,'MobilComSuper92000D1(Akquise)',NULL,NULL,'MS9ND1',327,24,'MobilCom Intern',7003,NULL,'auto',20010202105916,'Mobilfunk','PP','','',''); +SELECT ELT(FIELD(kundentyp,'PP','PPA','PG','PGA','FK','FKA','FP','FPA','K','KA','V','VA',''), 'Privat (Private Nutzung)','Privat (Private Nutzung) Sitz im Ausland','Privat (geschaeftliche Nutzung)','Privat (geschaeftliche Nutzung) Sitz im Ausland','Firma (Kapitalgesellschaft)','Firma (Kapitalgesellschaft) Sitz im Ausland','Firma (Personengesellschaft)','Firma (Personengesellschaft) Sitz im Ausland','oeff. rechtl. Koerperschaft','oeff. rechtl. Koerperschaft Sitz im Ausland','Eingetragener Verein','Eingetragener Verein Sitz im Ausland','Typ unbekannt') AS Kundentyp ,kategorie FROM t1 WHERE hdl_nr < 2000000 AND kategorie IN ('Prepaid','Mobilfunk') AND st_klasse = 'Workflow' GROUP BY kundentyp ORDER BY kategorie; +``` + +MySQL got: + +``` ++--------------------------+-----------+ +| Kundentyp | kategorie | ++--------------------------+-----------+ +| Privat (Private Nutzung) | Mobilfunk | ++--------------------------+-----------+ +1 row in set, 1 warning (0.00 sec) + +mysql> show warnings; ++---------+------+----------------------------------------------------+ +| Level | Code | Message | ++---------+------+----------------------------------------------------+ +| Warning | 1052 | Column 'kundentyp' in group statement is ambiguous | ++---------+------+----------------------------------------------------+ +1 row in set (0.00 sec) +``` + +but TiDB got empty set; + + + ",1,group by with alias variable not consistent with mysql test sql drop table if exists set sql mode no engine substitution create table cont nr int not null auto increment ver nr int not null default aufnr int not null default username varchar not null default hdl nr int not null default eintrag date not null default st klasse varchar not null default st wert varchar not null default st zusatz varchar not null default st bemerkung varchar not null default kunden art varchar not null default mcbs knr int default null mcbs aufnr int not null default schufa status char default bemerkung text wirknetz text wf igz int not null default tarifcode varchar default null recycle char default null sim varchar default null mcbs tpl varchar default null emp nr int not null default laufzeit int default null hdl name varchar default null prov hdl nr int not null default auto wirknetz varchar default null auto billing varchar default null touch timestamp not null kategorie varchar default null kundentyp varchar not null default sammel rech msisdn varchar not null default p nr varchar not null default suffix char not null default primary key cont nr key idx aufnr aufnr key idx hdl nr hdl nr key idx st klasse st klasse key ver nr ver nr key eintrag idx eintrag key emp nr idx emp nr key wf igz wf igz key touch touch key hdl tag eintrag hdl nr key prov hdl nr prov hdl nr key mcbs aufnr mcbs aufnr key kundentyp kundentyp key p nr p nr suffix insert into values mustermann musterfrau workflow auftrag erledigt originalvertrag eingegangen und gepr�ft privat n null null mobilcom shop koeln null auto mobilfunk pp insert into values mustermann musterfrau workflow auftrag erledigt originalvertrag eingegangen und gepr�ft privat p akquise null null mobilcom intern null auto mobilfunk pp insert into values mustermann musterfrau workflow auftrag erledigt originalvertrag eingegangen und gepr�ft privat n n null null mobilcom intern null auto mobilfunk pp insert into values mustermann musterfrau workflow auftrag erledigt originalvertrag eingegangen und gepr�ft privat p akquise null null mobilcom intern null auto mobilfunk pp insert into values mustermann musterfrau workflow auftrag erledigt originalvertrag eingegangen und gepr�ft privat null n null null null pos null null mobilfunk insert into values mustermann musterfrau workflow auftrag storniert storno kd m�chte nicht mehr privat null p akquise null null mobilcom intern null auto mobilfunk pp insert into values mustermann musterfrau workflow auftrag erledigt originalvertrag eingegangen und gepr�ft privat p akquise null null mobilcom intern null auto mobilfunk pp select elt field kundentyp pp ppa pg pga fk fka fp fpa k ka v va privat private nutzung privat private nutzung sitz im ausland privat geschaeftliche nutzung privat geschaeftliche nutzung sitz im ausland firma kapitalgesellschaft firma kapitalgesellschaft sitz im ausland firma personengesellschaft firma personengesellschaft sitz im ausland oeff rechtl koerperschaft oeff rechtl koerperschaft sitz im ausland eingetragener verein eingetragener verein sitz im ausland typ unbekannt as kundentyp kategorie from where hdl nr and kategorie in prepaid mobilfunk and st klasse workflow group by kundentyp order by kategorie mysql got kundentyp kategorie privat private nutzung mobilfunk row in set warning sec mysql show warnings level code message warning column kundentyp in group statement is ambiguous row in set sec but tidb got empty set ,1 65493,14727876424.0,IssuesEvent,2021-01-06 09:11:07,Seagate/cortx-s3server,https://api.github.com/repos/Seagate/cortx-s3server,closed,CVE-2018-16468 (Medium) detected in loofah-2.0.3.gem,needs-attention needs-triage security vulnerability,"## CVE-2018-16468 - Medium Severity VulnerabilityJavaScript library for DOM operations
+Library home page: https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.2/jquery.min.js
+Path to vulnerable library: /app/src/main/webapp/resources/js/jquery-1.11.2.min.js
+-``` -+---------------------------------------------------------------------------------------------------------------------------------------------------------+ -| -0.0000000000000000000000000000000000000000000000000017382578996420603 * -13890436710184412000000000000000000000000000000000000000000000000000000000000 | -+---------------------------------------------------------------------------------------------------------------------------------------------------------+ -| 0.000000000000000000000000000000 | -+---------------------------------------------------------------------------------------------------------------------------------------------------------+ +Dependency Hierarchy: + - :x: **jquery-1.11.2.min.js** (Vulnerable Library) +
Found in HEAD commit: 919af4911488ae49a0463c8d639240781e4088b0
+Found in base branch: main
+ ++ +jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. -``` +
Publish Date: 2019-04-20 +
URL: CVE-2019-11358
+ ++Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: Required + - Scope: Changed +- Impact Metrics: + - Confidentiality Impact: Low + - Integrity Impact: Low + - Availability Impact: None +
+For more information on CVSS3 Scores, click here. + +-3. What did you see instead? +
Type: Upgrade version
+Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358
+Release Date: 2019-04-20
+Fix Resolution: 3.4.0
-``` -ERROR 2006 (HY000): MySQL server has gone away -No connection. Trying to reconnect... -``` + +JavaScript library for DOM operations
+Library home page: https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.2/jquery.min.js
+Path to vulnerable library: /app/src/main/webapp/resources/js/jquery-1.11.2.min.js
+-``` -Release Version: v2.1.0-beta-11-g1f6e12cc5 -Git Commit Hash: 1f6e12cc55e534bf514eef212f1091a91b8592c9 -Git Branch: HEAD -UTC Build Time: 2018-07-04 06:31:26 -GoVersion: go version go1.10.1 linux/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -``` -",1,panic when multiply very big decimal with a very small decimal what did you do select what did you expect to see what did you see instead error mysql server has gone away no connection trying to reconnect conn go lastcmd select runtime error index out of range goroutine github com pingcap tidb server clientconn run home robi code go src github com pingcap tidb server conn go panic home robi src runtime panic go github com pingcap tidb types decimalmul home robi code go src github com pingcap tidb types mydecimal go github com pingcap tidb expression builtinarithmeticmultiplydecimalsig evaldecimal home robi code go src github com pingcap tidb expression builtin arithmetic go github com pingcap tidb expression scalarfunction evaldecimal home robi code go src github com pingcap tidb expression scalar function go github com pingcap tidb expression scalarfunction eval home robi code go src github com pingcap tidb expression scalar function go github com pingcap tidb expression foldconstant home robi code go src github com pingcap tidb expression constant fold go github com pingcap tidb expression foldconstant home robi code go src github com pingcap tidb expression constant fold go github com pingcap tidb expression newfunction home robi code go src github com pingcap tidb expression scalar function go github com pingcap tidb plan expressionrewriter binaryoptoexpression home robi code go src github com pingcap tidb plan expression rewriter go github com pingcap tidb plan expressionrewriter leave home robi code go src github com pingcap tidb plan expression rewriter go github com pingcap tidb ast binaryoperationexpr accept home robi code go src github com pingcap tidb ast expressions go github com pingcap tidb plan planbuilder rewritewithpreprocess home robi code go src github com pingcap tidb plan expression rewriter go github com pingcap tidb plan planbuilder rewrite home robi code go src github com pingcap tidb plan expression rewriter go github com pingcap tidb plan planbuilder buildprojection home robi code go src github com pingcap tidb plan logical plan builder go github com pingcap tidb plan planbuilder buildselect home robi code go src github com pingcap tidb plan logical plan builder go github com pingcap tidb plan planbuilder build home robi code go src github com pingcap tidb plan planbuilder go github com pingcap tidb plan optimize home robi code go src github com pingcap tidb plan optimizer go github com pingcap tidb executor compiler compile home robi code go src github com pingcap tidb executor compiler go github com pingcap tidb session session execute home robi code go src github com pingcap tidb session session go what version of tidb are you using tidb server v or run select tidb version on tidb release version beta git commit hash git branch head utc build time goversion go version linux race enabled false tikv min version alpha ,1 -272008,8494701283.0,IssuesEvent,2018-10-29 00:00:03,containous/traefik,https://api.github.com/repos/containous/traefik,closed,Access log records the wrong frontend name,contributor/waiting-for-feedback kind/bug/possible priority/P2,"### Do you want to request a *feature* or report a *bug*? +Dependency Hierarchy: + - :x: **jquery-1.11.2.min.js** (Vulnerable Library) +
Found in HEAD commit: 919af4911488ae49a0463c8d639240781e4088b0
+Found in base branch: main
+ ++ +jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. -Bug +
Publish Date: 2019-04-20 +
URL: CVE-2019-11358
+ +-### What did you do? +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: Required + - Scope: Changed +- Impact Metrics: + - Confidentiality Impact: Low + - Integrity Impact: Low + - Availability Impact: None +
+For more information on CVSS3 Scores, click here. + +-1. Using the consul config provider, create several frontends (with different Host rules) and point them to the same backend. -2. Turn on the `accessLog` feature. -3. Make a HTTP request to one of the frontend domains. +
Type: Upgrade version
+Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358
+Release Date: 2019-04-20
+Fix Resolution: 3.4.0
-### What did you expect to see? + +-### What did you see instead? +
Julia Cartwright's fork of linux-stable-rt.git
+Library home page: https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git
+Found in HEAD commit: 9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69
-The log message always seems to show the first frontend (alphabetically) from consul. +Found in base branch: master
+
-### Output of `traefik version`: (_What version of Traefik are you using?_)
+ /drivers/staging/rtl8188eu/os_dep/ioctl_linux.c
+
/drivers/staging/rtl8188eu/os_dep/ioctl_linux.c
+
+ +rtw_wx_set_scan in drivers/staging/rtl8188eu/os_dep/ioctl_linux.c in the Linux kernel through 5.11.6 allows writing beyond the end of the ->ssid[] array. NOTE: from the perspective of kernel.org releases, CVE IDs are not normally used for drivers/staging/* (unfinished work); however, system integrators may have situations in which a drivers/staging issue is relevant to their own customer base. -1.6.4 -",1.0,"Access log records the wrong frontend name - ### Do you want to request a *feature* or report a *bug*? +
Publish Date: 2021-03-17 +
URL: CVE-2021-28660
+ +-Bug +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Local + - Attack Complexity: Low + - Privileges Required: Low + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-### What did you do? +
Type: Upgrade version
+Origin: https://www.linuxkernelcves.com/cves/CVE-2021-28660
+Release Date: 2021-03-17
+Fix Resolution: v5.11.6
-1. Using the consul config provider, create several frontends (with different Host rules) and point them to the same backend. -2. Turn on the `accessLog` feature. -3. Make a HTTP request to one of the frontend domains. + +-The log message includes the name of the frontend matching the domain of the HTTP request. +
Julia Cartwright's fork of linux-stable-rt.git
+Library home page: https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git
+Found in HEAD commit: 9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69
-### What did you see instead? +Found in base branch: master
+
-The log message always seems to show the first frontend (alphabetically) from consul.
+ /drivers/staging/rtl8188eu/os_dep/ioctl_linux.c
+
/drivers/staging/rtl8188eu/os_dep/ioctl_linux.c
+
+ +rtw_wx_set_scan in drivers/staging/rtl8188eu/os_dep/ioctl_linux.c in the Linux kernel through 5.11.6 allows writing beyond the end of the ->ssid[] array. NOTE: from the perspective of kernel.org releases, CVE IDs are not normally used for drivers/staging/* (unfinished work); however, system integrators may have situations in which a drivers/staging issue is relevant to their own customer base. -### Output of `traefik version`: (_What version of Traefik are you using?_) +
Publish Date: 2021-03-17 +
URL: CVE-2021-28660
+ +-1.6.4 -",0,access log records the wrong frontend name do you want to request a feature or report a bug bug what did you do using the consul config provider create several frontends with different host rules and point them to the same backend turn on the accesslog feature make a http request to one of the frontend domains what did you expect to see the log message includes the name of the frontend matching the domain of the http request what did you see instead the log message always seems to show the first frontend alphabetically from consul output of traefik version what version of traefik are you using ,0 -7034,9307518771.0,IssuesEvent,2019-03-25 12:30:39,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,`SELECT 9223372036854775807 + 1;` is not compatible with MySQL,type/compatibility,"## Bug Report +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Local + - Attack Complexity: Low + - Privileges Required: Low + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-Please answer these questions before submitting your issue. Thanks! +
Type: Upgrade version
+Origin: https://www.linuxkernelcves.com/cves/CVE-2021-28660
+Release Date: 2021-03-17
+Fix Resolution: v5.11.6
-1. What did you do? -If possible, provide a recipe for reproducing the error. -`SELECT 9223372036854775807 + 1;` - -2. What did you expect to see? -In Mysql 8.0.15 -``` -mysql> SELECT 9223372036854775807+1; -ERROR 1690 (22003): BIGINT value is out of range in '(9223372036854775807 + 1)' -``` -See also https://dev.mysql.com/doc/refman/8.0/en/out-of-range-and-overflow.html - -3. What did you see instead? -``` -mysql> SELECT 9223372036854775807+1; -+-----------------------+ -| 9223372036854775807+1 | -+-----------------------+ -| 9223372036854775808 | -+-----------------------+ -1 row in set (0.00 sec) -``` + +-",True,"`SELECT 9223372036854775807 + 1;` is not compatible with MySQL - ## Bug Report - -Please answer these questions before submitting your issue. Thanks! +See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) +
++See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) +
+-4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? -``` -mysql> select tidb_version(); -+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| tidb_version() | -+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| Release Version: v3.0.0-beta-231-g20463d6-dirty -Git Commit Hash: 20463d6da90fdf12c0d9d18c15dc33a78334882d -Git Branch: master -UTC Build Time: 2019-03-20 02:03:04 -GoVersion: go version go1.12 linux/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -Check Table Before Drop: false | -+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -1 row in set (0.01 sec) -``` +
Julia Cartwright's fork of linux-stable-rt.git
+Library home page: https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git
+Found in HEAD commit: 5e23b7f9d2dd0154edd54986754eecd5b5308571
-",1, select is not compatible with mysql bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error select what did you expect to see in mysql mysql select error bigint value is out of range in see also what did you see instead mysql select row in set sec what version of tidb are you using tidb server v or run select tidb version on tidb mysql select tidb version tidb version release version beta dirty git commit hash git branch master utc build time goversion go version linux race enabled false tikv min version alpha check table before drop false row in set sec ,1 -6587,8854545270.0,IssuesEvent,2019-01-09 01:51:28,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,tidb-server options -socket and -host are incompatible,type/compatibility,"## Bug Report +Found in base branch: master
+
-Please answer these questions before submitting your issue. Thanks!
+ /drivers/net/wireless/ath/ath9k/xmit.c
+
+ +Race condition in the ath_tx_aggr_sleep function in drivers/net/wireless/ath/ath9k/xmit.c in the Linux kernel before 3.13.7 allows remote attackers to cause a denial of service (system crash) via a large amount of network traffic that triggers certain list deletions. -1. What did you do? +
Publish Date: 2014-04-01 +
URL: CVE-2014-2672
+ +-``` -./bin/tidb-server -socket /tmp/mysql.sock -host 127.0.0.1 -P 3306 -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: High + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-2. What did you expect to see? +
Type: Upgrade version
+Origin: https://nvd.nist.gov/vuln/detail/CVE-2014-2672
+Release Date: 2014-04-01
+Fix Resolution: 3.13.7
-I would expect it to listen on both. This is important because of MySQL's special handling of localhost: -- For the libmysql derived clients it will switch to using unix socket. -- For other drivers like JDBC it will only ever use tcp. + +-3. What did you see instead? +
Julia Cartwright's fork of linux-stable-rt.git
+Library home page: https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git
+Found in HEAD commit: 5e23b7f9d2dd0154edd54986754eecd5b5308571
- It will only listen on one (the socket!) +Found in base branch: master
+
-4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
+ /drivers/net/wireless/ath/ath9k/xmit.c
+
+ +Race condition in the ath_tx_aggr_sleep function in drivers/net/wireless/ath/ath9k/xmit.c in the Linux kernel before 3.13.7 allows remote attackers to cause a denial of service (system crash) via a large amount of network traffic that triggers certain list deletions. -``` -mysql> select tidb_version()\G -*************************** 1. row *************************** -tidb_version(): Release Version: v2.1.0-beta-644-g89cd59d21 -Git Commit Hash: 89cd59d217637f5e79fdc81681c98531c668e688 -Git Branch: newmaster -UTC Build Time: 2018-11-26 04:47:59 -GoVersion: go version go1.11 linux/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -Check Table Before Drop: false -1 row in set (0.00 sec) -```",True,"tidb-server options -socket and -host are incompatible - ## Bug Report - -Please answer these questions before submitting your issue. Thanks! - -1. What did you do? +
Publish Date: 2014-04-01 +
URL: CVE-2014-2672
+ +-``` -./bin/tidb-server -socket /tmp/mysql.sock -host 127.0.0.1 -P 3306 -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: High + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-2. What did you expect to see? +
Type: Upgrade version
+Origin: https://nvd.nist.gov/vuln/detail/CVE-2014-2672
+Release Date: 2014-04-01
+Fix Resolution: 3.13.7
-I would expect it to listen on both. This is important because of MySQL's special handling of localhost: -- For the libmysql derived clients it will switch to using unix socket. -- For other drivers like JDBC it will only ever use tcp. - -So it is useful to be able to have the server listen on both for a fully local install. + +Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.4.0/system.security.cryptography.xml.4.4.0.nupkg
+-4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? -``` -mysql> select tidb_version()\G -*************************** 1. row *************************** -tidb_version(): Release Version: v2.1.0-beta-644-g89cd59d21 -Git Commit Hash: 89cd59d217637f5e79fdc81681c98531c668e688 -Git Branch: newmaster -UTC Build Time: 2018-11-26 04:47:59 -GoVersion: go version go1.11 linux/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -Check Table Before Drop: false -1 row in set (0.00 sec) -```",1,tidb server options socket and host are incompatible bug report please answer these questions before submitting your issue thanks what did you do bin tidb server socket tmp mysql sock host p what did you expect to see i would expect it to listen on both this is important because of mysql s special handling of localhost for the libmysql derived clients it will switch to using unix socket for other drivers like jdbc it will only ever use tcp so it is useful to be able to have the server listen on both for a fully local install what did you see instead it will only listen on one the socket what version of tidb are you using tidb server v or run select tidb version on tidb mysql select tidb version g row tidb version release version beta git commit hash git branch newmaster utc build time goversion go version linux race enabled false tikv min version alpha check table before drop false row in set sec ,1 -51196,10595774186.0,IssuesEvent,2019-10-09 19:44:25,fdorg/flashdevelop,https://api.github.com/repos/fdorg/flashdevelop,opened,[Haxe][CodeGenerator] Add the generator `Generate constructor with initializer` for `abstract`,coderefactor enhancement haxe,"```haxe -abstract AType(Int) { - $(EntryPoint) -} -``` -expected result: -```haxe -abstract AType(Int) { - public function new(value:Int) this = value; -} -```",1.0,"[Haxe][CodeGenerator] Add the generator `Generate constructor with initializer` for `abstract` - ```haxe -abstract AType(Int) { - $(EntryPoint) -} -``` -expected result: -```haxe -abstract AType(Int) { - public function new(value:Int) this = value; -} -```",0, add the generator generate constructor with initializer for abstract haxe abstract atype int entrypoint expected result haxe abstract atype int public function new value int this value ,0 -31279,2732890259.0,IssuesEvent,2015-04-17 10:01:15,tiku01/oryx-editor,https://api.github.com/repos/tiku01/oryx-editor,closed,"Repository view: ""writer"" vs. ""contributor""",auto-migrated Priority-Medium Type-Enhancement,"``` -In the repository: On the left at ""access-filter"" it shows ""Me as writer"". -On the right (with selected model) you can add ""contributors"". For some -people this could be misleading. I prefer ""contributor"". -``` +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section ""Details"" below to see if there is a version of transitive dependency where vulnerability is fixed.
-Please answer these questions before submitting your issue. Thanks! +## Details -1. What did you do? -If possible, provide a recipe for reproducing the error. +-3. What did you see instead? -Got: `18446744073709550504` +### microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
-4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? +Core components of ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+-``` -+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| tidb_version() | -+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| Release Version: v2.1.8-24-g7b36cdfb6-dirty -Git Commit Hash: 7b36cdfb6e8a4c279fc360ed79cf1c499b553688 -Git Branch: failpoint-2.1 -UTC Build Time: 2019-04-30 08:26:17 -GoVersion: go version go1.12 darwin/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -Check Table Before Drop: false | -+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -``` -",True,"Incompatible with MySQL 8.0 about CAST string to unsigned - ## Bug Report +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) -Please answer these questions before submitting your issue. Thanks! -1. What did you do? -If possible, provide a recipe for reproducing the error. +### microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
-``` -MySQL(127.0.0.1:3306)> select cast('-1111.911' as unsigned) -+-------------------------------+ -| cast('-1111.911' as unsigned) | -+-------------------------------+ -| 18446744073709550505 | -+-------------------------------+ -1 row in set (0.002 sec) - -TiDB(127.0.0.1:4000)> select cast('-1111.911' as unsigned) -+-------------------------------+ -| cast('-1111.911' as unsigned) | -+-------------------------------+ -| 18446744073709550504 | -+-------------------------------+ -1 row in set (0.003 sec) -``` +ASP.NET Core basic middleware for supporting HTTP method overrides. Includes: +* X-Forwarded-* header...
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.httpoverrides/2.0.0/microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
+-2. What did you expect to see? +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg + - :x: **microsoft.aspnetcore.httpoverrides.2.0.0.nupkg** (Vulnerable Library) -Expected: `18446744073709550505` -3. What did you see instead? +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -Got: `18446744073709550504` + -4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? +### Vulnerability Details ++ +ASP.NET Core 1.0. 1.1, and 2.0 allow an elevation of privilege vulnerability due to how web applications that are created from templates validate web requests, aka ""ASP.NET Core Elevation Of Privilege Vulnerability"". -``` -+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| tidb_version() | -+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| Release Version: v2.1.8-24-g7b36cdfb6-dirty -Git Commit Hash: 7b36cdfb6e8a4c279fc360ed79cf1c499b553688 -Git Branch: failpoint-2.1 -UTC Build Time: 2019-04-30 08:26:17 -GoVersion: go version go1.12 darwin/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -Check Table Before Drop: false | -+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -``` -",1,incompatible with mysql about cast string to unsigned bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error mysql select cast as unsigned cast as unsigned row in set sec tidb select cast as unsigned cast as unsigned row in set sec what did you expect to see expected what did you see instead got what version of tidb are you using tidb server v or run select tidb version on tidb tidb version release version dirty git commit hash git branch failpoint utc build time goversion go version darwin race enabled false tikv min version alpha check table before drop false ,1 -16384,22092463572.0,IssuesEvent,2022-06-01 07:20:29,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,release savepoint behavior is inconsistent with mysql,type/bug severity/minor compatibility-mysql8,"## Bug Report - -Please answer these questions before submitting your issue. Thanks! - -### 1. Minimal reproduce step (Required) - -```sql -/* test */ select version(); -/* test */ drop table if exists t; -/* test */ create table t (k int primary key, v varchar(20)); -/* test */ begin; -/* test */ insert into t values (1, 'a'); -/* test */ savepoint p1; -/* test */ insert into t values (2, 'b'); -/* test */ savepoint p2; -/* test */ select * from t; -/* test */ release savepoint p1; -/* test */ select * from t; -/* test */ rollback to p2; -- tidb: OK; mysql: SAVEPOINT p2 does not exist; -/* test */ select * from t; -/* test */ rollback; -``` +
Publish Date: 2018-03-14 +
URL: CVE-2018-0787
+ + - +### CVSS 3 Score Details (8.8) +-### 2. What did you expect to see? (Required) +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: Required + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + -### 3. What did you see instead (Required) + -### 4. What is your TiDB version? (Required) +### Suggested Fix +- +
Type: Upgrade version
+Release Date: 2018-03-14
+Fix Resolution: Microsoft.AspNetCore.HttpOverrides - 2.0.2, Microsoft.AspNetCore.Server.Kestrel.Core - 2.0.2
-master: 2c3f717af66c6925f46776da867790ca150ecfcc + -",True,"release savepoint behavior is inconsistent with mysql - ## Bug Report + -Please answer these questions before submitting your issue. Thanks! +- -### 2. What did you expect to see? (Required) +### microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
-### 3. What did you see instead (Required) +Core components of ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+-### 4. What is your TiDB version? (Required) +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) - -master: 2c3f717af66c6925f46776da867790ca150ecfcc +### microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
-",1,release savepoint behavior is inconsistent with mysql bug report please answer these questions before submitting your issue thanks minimal reproduce step required sql test select version test drop table if exists t test create table t k int primary key v varchar test begin test insert into t values a test savepoint test insert into t values b test savepoint test select from t test release savepoint test select from t test rollback to tidb ok mysql savepoint does not exist test select from t test rollback what did you expect to see required what did you see instead required what is your tidb version required master ,1 -22492,6254129005.0,IssuesEvent,2017-07-14 00:41:22,diaspora/diaspora,https://api.github.com/repos/diaspora/diaspora,closed,"Message has redundant ""conversation_guid"" in it XML representation",code refactor federation,"In its federated XML representation Message model have property [conversation_guid](https://github.com/diaspora/diaspora/blob/develop/app/models/message.rb#L10) and property [parent_guid](https://github.com/diaspora/diaspora/blob/develop/lib/diaspora/relayable.rb#L12) from the Relayable inclusion. Looks like these properties duplicate each other. I think the former could be removed in favour to the latter. However it could ruin signature verifications, so it's not possible, before the new method of signature computation is adopted for the whole federation. -##Transport abstractions for the ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.transport.abstractions/2.0.0/microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
+-Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/29123253-message-has-redundant-conversation_guid-in-it-xml-representation?utm_campaign=plugin&utm_content=tracker%2F522&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F522&utm_medium=issues&utm_source=github). -
Libuv transport for the ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.transport.libuv/2.0.0/microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg
+-```mysql -SELECT NTH_VALUE(fieldA, -1) OVER (w1 PARTITION BY fieldB ORDER BY fieldB , fieldA ) AS 'ntile', fieldA, fieldB FROM ( SELECT `pk` AS fieldA, `col_date_key` AS fieldB FROM `table70_int_autoinc` ) as t WINDOW w1 AS ( ORDER BY fieldB ASC, fieldA DESC ); -``` +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg** (Vulnerable Library) -2. What did you expect to see? -```mysql -ERROR 1210 (HY000): Incorrect arguments to nth_value -``` +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -3. What did you see instead? + -```mysql -ERROR 3581 (HY000): A window which depends on another cannot define partitioning. -``` -4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? +### Vulnerability Details ++ +A vulnerability was discovered in versions 2.x of ASP.NET Core where a specially crafted request can cause excess resource consumption in Kestrel. -",True,"window function: should check fuction's arguments firstly. - ## Bug Report +
Publish Date: 2018-05-08 +
URL: WS-2018-0608
+ -Please answer these questions before submitting your issue. Thanks! + -1. What did you do? -If possible, provide a recipe for reproducing the error. +### CVSS 3 Score Details (7.5) +-```mysql -SELECT NTH_VALUE(fieldA, -1) OVER (w1 PARTITION BY fieldB ORDER BY fieldB , fieldA ) AS 'ntile', fieldA, fieldB FROM ( SELECT `pk` AS fieldA, `col_date_key` AS fieldB FROM `table70_int_autoinc` ) as t WINDOW w1 AS ( ORDER BY fieldB ASC, fieldA DESC ); -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + -2. What did you expect to see? + -```mysql -ERROR 1210 (HY000): Incorrect arguments to nth_value -``` +### Suggested Fix +-3. What did you see instead? +
Type: Upgrade version
+Release Date: 2018-05-08
+Fix Resolution: Microsoft.AspNetCore.Server.Kestrel.Core - 2.0.3,2.1.0;Microsoft.AspNetCore.Server.Kestrel.Transport.Abstractions - 2.0.3,2.1.0;Microsoft.AspNetCore.Server.Kestrel.Transport.Libuv - 2.0.3,2.1.0;Microsoft.AspNetCore.All - 2.0.8,2.1.0
-```mysql -ERROR 3581 (HY000): A window which depends on another cannot define partitioning. -``` -4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? + -",1,window function should check fuction s arguments firstly bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error mysql select nth value fielda over partition by fieldb order by fieldb fielda as ntile fielda fieldb from select pk as fielda col date key as fieldb from int autoinc as t window as order by fieldb asc fielda desc what did you expect to see mysql error incorrect arguments to nth value what did you see instead mysql error a window which depends on another cannot define partitioning what version of tidb are you using tidb server v or run select tidb version on tidb ,1 -309329,23291607790.0,IssuesEvent,2022-08-06 00:30:36,TavaresDylan/ICG,https://api.github.com/repos/TavaresDylan/ICG,closed,fix(doc): mkdocs error page no css,bug scope/documentation stale,"**Bug description** -404 page of mkdocs is show without css + -**Steps to reproduce** -- [ ] Just go on non existant page in documentation +-1. What did you do? -```sql -SELECT CHARSET('a'); -``` +### microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg
-2. What did you expect to see? +ASP.NET Core components for working with the IIS AspNetCoreModule.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.iisintegration/2.0.0/microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg
+-``` -+--------------+ -| CHARSET('a') | -+--------------+ -| utf8 | -+--------------+ -1 row in set (0.00 sec) -``` +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg** (Vulnerable Library) -3. What did you see instead? -``` -ERROR 1305 (42000): FUNCTION CHARSET does not exist -``` +### microsoft.aspnetcore.hosting.2.0.0.nupkg
-4. What version of TiDB are you using (`tidb-server -V`)? +ASP.NET Core hosting infrastructure and startup logic for web applications.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.hosting.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.hosting/2.0.0/microsoft.aspnetcore.hosting.2.0.0.nupkg
+-",True,"Built-in function CHARSET not supported - Please answer these questions before submitting your issue. Thanks! +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.hosting.2.0.0.nupkg** (Vulnerable Library) -1. What did you do? -```sql -SELECT CHARSET('a'); -``` +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -2. What did you expect to see? + -``` -+--------------+ -| CHARSET('a') | -+--------------+ -| utf8 | -+--------------+ -1 row in set (0.00 sec) -``` - -3. What did you see instead? - -``` -ERROR 1305 (42000): FUNCTION CHARSET does not exist -``` - -4. What version of TiDB are you using (`tidb-server -V`)? - -",1,built in function charset not supported please answer these questions before submitting your issue thanks what did you do sql select charset a what did you expect to see charset a row in set sec what did you see instead error function charset does not exist what version of tidb are you using tidb server v ,1 -25651,12701403801.0,IssuesEvent,2020-06-22 18:06:11,bbc/simorgh,https://api.github.com/repos/bbc/simorgh,opened,Serve images in next-gen formats,Refinement Needed performance ws-stories,"**Is your feature request related to a problem? Please describe.** -Parent issue: https://github.com/bbc/simorgh-infrastructure/issues/1088 - -We serve images in JPG format, but formats like JPEG 200, JPEG XR and WebP often provide better compression, which means faster downloads and less data consumption. - -**Describe the solution you'd like** -Serve images in next-gen formats. +### Vulnerability Details ++ +ASP.NET Core 1.0. 1.1, and 2.0 allow an elevation of privilege vulnerability due to how ASP.NET web applications handle web requests, aka ""ASP.NET Core Elevation Of Privilege Vulnerability"". This CVE is unique from CVE-2018-0784. -More information: https://web.dev/uses-webp-images/ +
Publish Date: 2018-03-14 +
URL: CVE-2018-0808
+ -**Describe alternatives you've considered** -A clear and concise description of any alternative solutions or features you've considered. + -**Testing notes** -[Tester to complete] +### CVSS 3 Score Details (7.5) +-Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + -- [ ] This feature is expected to need manual testing. + -**Additional context** -Add any other context or screenshots about the feature request here. -",True,"Serve images in next-gen formats - **Is your feature request related to a problem? Please describe.** -Parent issue: https://github.com/bbc/simorgh-infrastructure/issues/1088 +### Suggested Fix +-We serve images in JPG format, but formats like JPEG 200, JPEG XR and WebP often provide better compression, which means faster downloads and less data consumption. +
Type: Upgrade version
+Origin: https://portal.msrc.microsoft.com/en-US/security-guidance/advisory/CVE-2018-0808
+Release Date: 2018-03-14
+Fix Resolution: Microsoft.AspNetCore.Server.IISIntegration - 2.1.0, Microsoft.AspNetCore.Hosting - 2.1.0
-**Describe the solution you'd like** -Serve images in next-gen formats. + -More information: https://web.dev/uses-webp-images/ + -**Describe alternatives you've considered** -A clear and concise description of any alternative solutions or features you've considered. +Core components of ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+-Please answer these questions before submitting your issue. Thanks! +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -1. What did you do? + -``` -mysql> select uncompress(unhex('02000000789CCB48CDC9C95728CF2FCA4901001A0B045D')); -``` +### Vulnerability Details ++ +Denial of service vulnerability in ASP.NET Core when a malformed request is terminated. -2. What did you expect to see? - -``` -mysql> select uncompress(unhex('02000000789CCB48CDC9C95728CF2FCA4901001A0B045D')); -+---------------------------------------------------------------------+ -| uncompress(unhex('02000000789CCB48CDC9C95728CF2FCA4901001A0B045D')) | -+---------------------------------------------------------------------+ -| NULL | -+---------------------------------------------------------------------+ -1 row in set, 1 warning (0.00 sec) - -mysql> show warnings; -+---------+------+--------------------------------------------------------------------------------------------------+ -| Level | Code | Message | -+---------+------+--------------------------------------------------------------------------------------------------+ -| Warning | 1258 | ZLIB: Not enough room in the output buffer (probably, length of uncompressed data was corrupted) | -+---------+------+--------------------------------------------------------------------------------------------------+ -1 row in set (0.00 sec) -``` +
Publish Date: 2018-07-10 +
URL: WS-2018-0607
+ -3. What did you see instead? -``` -+---------------------------------------------------------------------+ -| uncompress(unhex('02000000789CCB48CDC9C95728CF2FCA4901001A0B045D')) | -+---------------------------------------------------------------------+ -| hello world | -+---------------------------------------------------------------------+ -1 row in set (0.01 sec) -``` + -4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? +### CVSS 3 Score Details (7.5) +-``` -tidb_version(): Release Version: v2.1.0-rc.2-10-gb30dbd0 -Git Commit Hash: b30dbd06acaab088a1e8fa004e189c651ca08bf1 -Git Branch: master -UTC Build Time: 2018-09-17 01:20:48 -GoVersion: go version go1.11 linux/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -Check Table Before Drop: false -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + -5. more info -This is found when I try to implement builtin compress and uncompress function in TiKV (https://github.com/tikv/tikv/pull/3856) -the first four bytes stored the length of the uncompress string, if it's less than the corrected length (maybe corrupted), MySQL return Null, TiDB just ignore it.",True,"builtin uncompress function ignored the four bytes length - ## Bug Report + -Please answer these questions before submitting your issue. Thanks! +### Suggested Fix +-1. What did you do? +
Type: Upgrade version
+Release Date: 2018-07-10
+Fix Resolution: Microsoft.AspNetCore.Server.Kestrel.Core - 2.1.2
-``` -mysql> select uncompress(unhex('02000000789CCB48CDC9C95728CF2FCA4901001A0B045D')); -``` + -2. What did you expect to see? - -``` -mysql> select uncompress(unhex('02000000789CCB48CDC9C95728CF2FCA4901001A0B045D')); -+---------------------------------------------------------------------+ -| uncompress(unhex('02000000789CCB48CDC9C95728CF2FCA4901001A0B045D')) | -+---------------------------------------------------------------------+ -| NULL | -+---------------------------------------------------------------------+ -1 row in set, 1 warning (0.00 sec) - -mysql> show warnings; -+---------+------+--------------------------------------------------------------------------------------------------+ -| Level | Code | Message | -+---------+------+--------------------------------------------------------------------------------------------------+ -| Warning | 1258 | ZLIB: Not enough room in the output buffer (probably, length of uncompressed data was corrupted) | -+---------+------+--------------------------------------------------------------------------------------------------+ -1 row in set (0.00 sec) -``` + -3. What did you see instead? -``` -+---------------------------------------------------------------------+ -| uncompress(unhex('02000000789CCB48CDC9C95728CF2FCA4901001A0B045D')) | -+---------------------------------------------------------------------+ -| hello world | -+---------------------------------------------------------------------+ -1 row in set (0.01 sec) -``` +ASP.NET Core default HTTP feature implementations.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.http.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.http/2.0.0/microsoft.aspnetcore.http.2.0.0.nupkg
+-### Example +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - microsoft.aspnetcore.hosting.2.0.0.nupkg + - :x: **microsoft.aspnetcore.http.2.0.0.nupkg** (Vulnerable Library) +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -The [`TABLE`](https://dev.mysql.com/doc/refman/8.0/en/table.html) and [`VALUES`](https://dev.mysql.com/doc/refman/8.0/en/values.html) statements are two query specification statements introduced in [MySQL 8.0.19](https://dev.mysql.com/doc/relnotes/mysql/8.0/en/news-8-0-19.html#mysqld-8-0-19-sql-syntax). + -```sql -TABLE tbl_name [ORDER BY col_name] [LIMIT limit [OFFSET offset]]; +### Vulnerability Details ++ +A security feature bypass vulnerability exists in the way Microsoft ASP.NET Core parses encoded cookie names.The ASP.NET Core cookie parser decodes entire cookie strings which could allow a malicious attacker to set a second cookie with the name being percent encoded.The security update addresses the vulnerability by fixing the way the ASP.NET Core cookie parser handles encoded names., aka 'Microsoft ASP.NET Core Security Feature Bypass Vulnerability'. -VALUES ROW(val1, val2, …), … [ORDER BY column_N [LIMIT limit [OFFSET offset]]]; -``` +
Publish Date: 2020-09-11 +
URL: CVE-2020-1045
+ -### Explaination + -These can be used in place everywhere a `SELECT` statement was expected, including: +### CVSS 3 Score Details (7.5) +-* UNION -* JOIN -* subqueries -* SELECT INTO OUTFILE -* INSERT INTO SELECT -* CREATE TABLE SELECT +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: High + - Availability Impact: None +
+For more information on CVSS3 Scores, click here. + + -### Solution -For a reference, please see https://github.com/pingcap/parser/pull/680 and learn how to support a new syntax in the parser. +### Suggested Fix +-## Score +
Type: Upgrade version
+Release Date: 2020-09-11
+Fix Resolution: Microsoft.AspNetCore.App - 2.1.22, Microsoft.AspNetCore.All - 2.1.22,Microsoft.NETCore.App - 2.1.22, Microsoft.AspNetCore.Http - 2.1.22
-* 300 + -## SIG Slack Channel + -You can join [#sig-ddl](https://app.slack.com/client/TH91JCS4W/CMAKWBNJU) on slack in your spare time to discuss and get help with mentors or others. +Core components of ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+-* Parser -* Golang +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -## Learning Materials + -* [Parser Quickstart](https://github.com/pingcap/parser/blob/master/docs/quickstart.md) -* [Parser implementation(CN)](https://pingcap.com/blog-cn/tidb-source-code-reading-5/) -",True,"Support VALUES and TABLE statements syntax - ## Description +### Vulnerability Details ++ +ASP.NET Core and Visual Studio Denial of Service Vulnerability -Related parser issue: https://github.com/pingcap/parser/issues/869 +
Publish Date: 2021-01-12 +
URL: CVE-2021-1723
+ -### Example + -The [`TABLE`](https://dev.mysql.com/doc/refman/8.0/en/table.html) and [`VALUES`](https://dev.mysql.com/doc/refman/8.0/en/values.html) statements are two query specification statements introduced in [MySQL 8.0.19](https://dev.mysql.com/doc/relnotes/mysql/8.0/en/news-8-0-19.html#mysqld-8-0-19-sql-syntax). +### CVSS 3 Score Details (7.5) +-```sql -TABLE tbl_name [ORDER BY col_name] [LIMIT limit [OFFSET offset]]; +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + -VALUES ROW(val1, val2, …), … [ORDER BY column_N [LIMIT limit [OFFSET offset]]]; -``` + -### Explaination +### Suggested Fix +-These can be used in place everywhere a `SELECT` statement was expected, including: +
Type: Upgrade version
+Origin: https://nvd.nist.gov/vuln/detail/CVE-2021-1723
+Release Date: 2021-01-12
+Fix Resolution: Microsoft.AspNetCore.App.Runtime.win-arm64 - 5.0.2;LiveReloadServer - 1.1.0;Plugga.Core - 1.0.2;Maple.Branch.Module - 1.0.4;Microsoft.AspNetCore.Components.WebAssembly.Server - 5.0.1,5.0.0-rc.1.20451.17;AspNetCoreRuntime.5.0.x64 - 5.0.2;AspNetCoreRuntime.5.0.x86 - 5.0.2;Microsoft.AspNetCore.App.Runtime.osx-x64 - 5.0.2,3.1.10;GrazeDocs - 2.0.1;Microsoft.AspNetCore.App.Runtime.linux-musl-arm - 5.0.2;Microsoft.AspNetCore.App.Runtime.linux-musl-x64 - 5.0.2,3.1.10;YHWins.Template - 1.1.0;Microsoft.AspNetCore.App.Runtime.linux-musl-arm64 - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Runtime.linux-arm64 - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Ref - 3.1.10,6.0.0-rc.1.21452.15;Microsoft.AspNetCore.Blazor.DevServer - 3.2.0-preview1.20073.1,3.1.0-preview4.19579.2;Microsoft.AspNetCore.App.Runtime.linux-arm - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Runtime.linux-x64 - 3.1.10,5.0.2;stankins.console - 2020.12.20-beta298;Toolbelt.Blazor.DevServer.WithCssLiveReloader - 5.0.1,5.0.0-rc.1.20451.17;DragonFire.Server - 0.0.1-alpha.0;PoExtractor.OrchardCore - 0.5.0-rc2-16220;Microsoft.AspNetCore.App.Runtime.win-arm - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Runtime.win-x64 - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Runtime.win-x86 - 3.1.10,5.0.2;HuLu.Template.Api - 1.0.2;AspNetCoreRuntime.3.1.x64 - 3.1.10;AspNetCoreRuntime.3.1.x86 - 3.1.10;Microsoft.AspNetCore.Components.WebAssembly.DevServer - 5.0.0-rc.1.20451.17,5.0.1;Microsoft.AspNetCore.App.Runtime.win-arm64 - 3.1.10;lingman-webapi - 0.0.18
-* UNION -* JOIN -* subqueries -* SELECT INTO OUTFILE -* INSERT INTO SELECT -* CREATE TABLE SELECT + + -### Solution -For a reference, please see https://github.com/pingcap/parser/pull/680 and learn how to support a new syntax in the parser. +Provides classes to support the creation and validation of XML digital signatures. The classes in th...
+Library home page: https://api.nuget.org/packages/system.security.cryptography.xml.4.4.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.4.0/system.security.cryptography.xml.4.4.0.nupkg
+-## Mentor(s) +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.mvc.2.0.0.nupkg + - microsoft.aspnetcore.mvc.viewfeatures.2.0.0.nupkg + - microsoft.aspnetcore.antiforgery.2.0.0.nupkg + - microsoft.aspnetcore.dataprotection.2.0.0.nupkg + - :x: **system.security.cryptography.xml.4.4.0.nupkg** (Vulnerable Library) +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -* @tangenta + -Contact the mentors: **#tidb-challenge-program** channel in [TiDB Community](https://join.slack.com/t/tidbcommunity/shared_invite/enQtNzc0MzI4ODExMDc4LWYwYmIzMjZkYzJiNDUxMmZlN2FiMGJkZjAyMzQ5NGU0NGY0NzI3NTYwMjAyNGQ1N2I2ZjAxNzc1OGUwYWM0NzE) Slack Workspace +### Vulnerability Details ++ +Microsoft is releasing this security advisory to provide information about a vulnerability in .NET Core 3.1 and .NET 6.0. An information disclosure vulnerability exists in .NET Core 3.1 and .NET 6.0 that could lead to unauthorized access of privileged information. -## Recommended Skills +## Affected software -* Parser -* Golang +* Any .NET 6.0 application running on .NET 6.0.7 or earlier. +* Any .NET Core 3.1 applicaiton running on .NET Core 3.1.27 or earlier. -## Learning Materials +## Patches -* [Parser Quickstart](https://github.com/pingcap/parser/blob/master/docs/quickstart.md) -* [Parser implementation(CN)](https://pingcap.com/blog-cn/tidb-source-code-reading-5/) -",1,support values and table statements syntax description related parser issue example the and statements are two query specification statements introduced in sql table tbl name values row … … explaination these can be used in place everywhere a select statement was expected including union join subqueries select into outfile insert into select create table select solution for a reference please see and learn how to support a new syntax in the parser score sig slack channel you can join on slack in your spare time to discuss and get help with mentors or others mentor s tangenta contact the mentors tidb challenge program channel in slack workspace recommended skills parser golang learning materials ,1 -27185,7922117851.0,IssuesEvent,2018-07-05 09:44:48,ShaikASK/Testing,https://api.github.com/repos/ShaikASK/Testing,closed,Safari(11) : New Hires : New Hire /Re Hire Header & Save /cancel buttons are missing due to missing of vertical scroll bar ,Defect HR Admin Module New Hire P1 Release #3 Build 8,"Steps to Replicate : +* If you're using .NET 6.0, you should download and install Runtime 6.0.8 or SDK 6.0.108 (for Visual Studio 2022 v17.1) from https://dotnet.microsoft.com/download/dotnet-core/6.0. +* If you're using .NET Core 3.1, you should download and install Runtime 3.1.28 (for Visual Studio 2019 v16.9) from https://dotnet.microsoft.com/download/dotnet-core/3.1. -1. Launch the URL -2. Sign in as HR Admin user -3. Create a New Hire by entering full length of text +
Publish Date: 2022-08-09 +
URL: CVE-2022-34716
+ -Experienced Behaviour : Observed that New Hire /Re Hire Header & Save /cancel buttons are missing upon entering full length text in Notes field Safari (11) -Expected Behaviour : Ensure that New Hire /Re Hire Header & Save /cancel buttons buttons should be displayed upon entering full length of text ",1.0,"Safari(11) : New Hires : New Hire /Re Hire Header & Save /cancel buttons are missing due to missing of vertical scroll bar - Steps to Replicate : + -1. Launch the URL -2. Sign in as HR Admin user -3. Create a New Hire by entering full length of text +### CVSS 3 Score Details (5.9) +-Experienced Behaviour : Observed that New Hire /Re Hire Header & Save /cancel buttons are missing upon entering full length text in Notes field Safari (11) -Expected Behaviour : Ensure that New Hire /Re Hire Header & Save /cancel buttons buttons should be displayed upon entering full length of text ",0,safari new hires new hire re hire header save cancel buttons are missing due to missing of vertical scroll bar steps to replicate launch the url sign in as hr admin user create a new hire by entering full length of text experienced behaviour observed that new hire re hire header save cancel buttons are missing upon entering full length text in notes field safari expected behaviour ensure that new hire re hire header save cancel buttons buttons should be displayed upon entering full length of text ,0 -176171,28039053342.0,IssuesEvent,2023-03-28 17:03:48,pulumi/pulumi-docker,https://api.github.com/repos/pulumi/pulumi-docker,closed,docker:index:Image could not open dockerfile at relative path Dockerfile,kind/bug resolution/by-design,"### What happened? +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: High + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: None + - Availability Impact: None +
+For more information on CVSS3 Scores, click here. + -Upgraded from `pulumi-docker` `4.0.0` to `4.1.0`. -`pulumi up` unexpectedly errors: + -``` -Diagnostics: - docker:index:Image (my-app:latest): - error: could not open dockerfile at relative path Dockerfile. Try setting `dockerfile` to ""/Users/shed/Repos/my-repo/projects/my-app/Dockerfile"" -``` +### Suggested Fix +-### Expected Behavior - -`pulumi up` to locate `Dockerfile` relative to my project directory. - -### Steps to reproduce - -My pulumi resource looks something like: - -```python -image = docker.Image( - ""my_app:latest"", - build=docker.DockerBuildArgs( - context=""/Users/shed/Repos/my-repo/projects/my-app/"", - dockerfile=""Dockerfile"", - args={""BUILDKIT_INLINE_CACHE"": ""1""}, # may not be necessary? - cache_from=docker.CacheFromArgs( - images=[ - pulumi.Output.concat(repo.repository_url, "":"", ""latest"") - ] - ), - platform=""linux/amd64"", - ), - registry=registry, - image_name=pulumi.Output.concat(self.repo.repository_url, "":"", ""latest""), - ) -``` +
Type: Upgrade version
+Origin: https://github.com/advisories/GHSA-2m65-m22p-9wjw
+Release Date: 2022-08-09
+Fix Resolution: Microsoft.AspNetCore.App.Runtime.linux-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.osx-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x86 - 3.1.28,6.0.8;System.Security.Cryptography.Xml - 4.7.1,6.0.1
-There is a file `/Users/shed/Repos/my-repo/projects/my-app/Dockerfile`. -`context` is set to the directory and `dockerfile` is explicitly set to `Dockerfile`. + + -### Output of `pulumi about` +Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.4.0/system.security.cryptography.xml.4.4.0.nupkg
+-### Additional context -Split from #566 +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section ""Details"" below to see if there is a version of transitive dependency where vulnerability is fixed.
-Upgraded from `pulumi-docker` `4.0.0` to `4.1.0`. -`pulumi up` unexpectedly errors: +## Details -``` -Diagnostics: - docker:index:Image (my-app:latest): - error: could not open dockerfile at relative path Dockerfile. Try setting `dockerfile` to ""/Users/shed/Repos/my-repo/projects/my-app/Dockerfile"" -``` +-### Output of `pulumi about` -I don't want to copy/paste this from a work project. I'll try to provide specific info as requested if relevant. +### microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
-### Additional context +Core components of ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+-Split from #566 +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) -### Contributing -Vote on this issue by adding a 👍 reaction. -To contribute a fix for this issue, leave a comment (and link to your pull request, if you've opened one already). -",0,docker index image could not open dockerfile at relative path dockerfile what happened upgraded from pulumi docker to pulumi up unexpectedly errors diagnostics docker index image my app latest error could not open dockerfile at relative path dockerfile try setting dockerfile to users shed repos my repo projects my app dockerfile expected behavior pulumi up to locate dockerfile relative to my project directory steps to reproduce my pulumi resource looks something like python image docker image my app latest build docker dockerbuildargs context users shed repos my repo projects my app dockerfile dockerfile args buildkit inline cache may not be necessary cache from docker cachefromargs images pulumi output concat repo repository url latest platform linux registry registry image name pulumi output concat self repo repository url latest there is a file users shed repos my repo projects my app dockerfile context is set to the directory and dockerfile is explicitly set to dockerfile output of pulumi about i don t want to copy paste this from a work project i ll try to provide specific info as requested if relevant additional context split from contributing vote on this issue by adding a 👍 reaction to contribute a fix for this issue leave a comment and link to your pull request if you ve opened one already ,0 -17959,10169183491.0,IssuesEvent,2019-08-07 23:19:28,dotnet/corefx,https://api.github.com/repos/dotnet/corefx,closed,ASN.1 Deserialize fails with SequenceOf strings,area-System.Security,"Let me preface this with the statement that I fully understand this is an internal-only API (for now at least). Nevertheless, this is a fairly important test case. +### microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
-Suppose an ASN.1 of +ASP.NET Core basic middleware for supporting HTTP method overrides. Includes: +* X-Forwarded-* header...
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.httpoverrides/2.0.0/microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
+-``` -kdc-req ::= SEQUENCE { -... - sname [3] PrincipalName OPTIONAL, -... -} +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg + - :x: **microsoft.aspnetcore.httpoverrides.2.0.0.nupkg** (Vulnerable Library) -PrincipalName ::= SEQUENCE { - name-type [0] Int32, - name-string [1] SEQUENCE OF KerberosString -} -KerberosString ::= GeneralString (IA5String) -``` +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -There are two issues here. + -1. GeneralString is not actually supported. You cannot futz your way through by decorating the struct field to think it's an IA5String. I had to add a new attribute and modified the deserializer to treat it the same as IA5String. -2. SequenceOf GeneralString/IA5String cannot parse correctly *when there are multiple values in the sequence*. +### Vulnerability Details ++ +ASP.NET Core 1.0. 1.1, and 2.0 allow an elevation of privilege vulnerability due to how web applications that are created from templates validate web requests, aka ""ASP.NET Core Elevation Of Privilege Vulnerability"". -It decodes the array as a struct and throws because there's extra data. There's no way it can move on to the next element in the sequence. Conversely, you can't coerce the deserializer to operate against string[] because it can't be a SequenceOf and GeneralizedString/IA5String at the same time. +
Publish Date: 2018-03-14 +
URL: CVE-2018-0787
+ + +### CVSS 3 Score Details (8.8) +-### Sample +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: Required + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + -This will throw when it finishes reading the first string of `sname`. It'll complete the pass over the struct fields, but there's more data in the sequence. + -I think it would be helpful if you could decorate a `SequenceOf` as a specialized type, e.g. `GeneralizedString`. +### Suggested Fix +-```c# -var asReq = ""3081d3a103020105a20302010aa31530133011a10402020080a20904073005a0030101ffa481af30"" -+ ""81aca00703050040810010a1153013a003020101a10c300a1b087374737966756873a2091b07524"" -+ ""5444d4f4e44a31c301aa003020102a11330111b066b72627467741b075245444d4f4e44a511180f3"" -+ ""2303337303931333032343830355aa611180f32303337303931333032343830355aa70502036875"" -+ ""0da81530130201120201110201170201180202ff79020103a91d301b3019a003020114a11204104"" -+ ""d494e494e542d364a49454842472020""; +
Type: Upgrade version
+Release Date: 2018-03-14
+Fix Resolution: Microsoft.AspNetCore.HttpOverrides - 2.0.2, Microsoft.AspNetCore.Server.Kestrel.Core - 2.0.2
-var req = AsnSerializer.Deserialize
-[StructLayout(LayoutKind.Sequential)]
-public struct KdcReqBody
-{
- /*
- kdc-options [0] KDCOptions,
- cname [1] PrincipalName OPTIONAL
- -- Used only in AS-REQ --,
- realm [2] Realm
- -- Server's realm
- -- Also client's in AS-REQ --,
- sname [3] PrincipalName OPTIONAL,
- from [4] KerberosTime OPTIONAL,
- till [5] KerberosTime,
- rtime [6] KerberosTime OPTIONAL,
- nonce [7] UInt32,
- etype [8] SEQUENCE OF Int32 -- EncryptionType
- -- in preference order --,
- addresses [9] HostAddresses OPTIONAL,
- enc-authorization-data [10] EncryptedData OPTIONAL
- -- AuthorizationData --,
- additional-tickets [11] SEQUENCE OF Ticket OPTIONAL
- -- NOTE: not empty
- */
- [ExpectedTag(0, ExplicitTag = true), BitString]
- public ReadOnlyMemory
Core components of ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+- [ExpectedTag(1)] - public KerberosString[] nameString; +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) - // this does not work either - // [ExpectedTag(1), GeneralizedString] - // public string[] nameString; -} -[StructLayout(LayoutKind.Sequential)] -public struct KerberosString -{ - [GeneralizedString] - public string name; -} +### microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
-[StructLayout(LayoutKind.Sequential)] -public struct HostAddress -{ - // addr-type[0] Int32, - // address[1] OCTET STRING +Transport abstractions for the ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.transport.abstractions/2.0.0/microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
+
- [ExpectedTag(0, ExplicitTag = true)]
- public int addrType;
+Dependency Hierarchy:
+ - dotnetcore-sample-1.0.0 (Root Library)
+ - microsoft.aspnetcore.2.0.0.nupkg
+ - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg
+ - microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg
+ - :x: **microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg** (Vulnerable Library)
- [ExpectedTag(1, ExplicitTag = true), OctetString]
- public ReadOnlyMemory
Libuv transport for the ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.transport.libuv/2.0.0/microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg
+
- [ExpectedTag(1, ExplicitTag = true), OptionalValue]
- public int kvno;
+Dependency Hierarchy:
+ - dotnetcore-sample-1.0.0 (Root Library)
+ - microsoft.aspnetcore.2.0.0.nupkg
+ - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg
+ - :x: **microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg** (Vulnerable Library)
- [ExpectedTag(2, ExplicitTag = true), OctetString]
- public ReadOnlyMemory Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9 Found in base branch: main
+ +A vulnerability was discovered in versions 2.x of ASP.NET Core where a specially crafted request can cause excess resource consumption in Kestrel. -Suppose an ASN.1 of +
Publish Date: 2018-05-08 +
URL: WS-2018-0608
+ -``` -kdc-req ::= SEQUENCE { -... - sname [3] PrincipalName OPTIONAL, -... -} + -PrincipalName ::= SEQUENCE { - name-type [0] Int32, - name-string [1] SEQUENCE OF KerberosString -} +### CVSS 3 Score Details (7.5) +-KerberosString ::= GeneralString (IA5String) -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + -There are two issues here. + -1. GeneralString is not actually supported. You cannot futz your way through by decorating the struct field to think it's an IA5String. I had to add a new attribute and modified the deserializer to treat it the same as IA5String. -2. SequenceOf GeneralString/IA5String cannot parse correctly *when there are multiple values in the sequence*. +### Suggested Fix +-It decodes the array as a struct and throws because there's extra data. There's no way it can move on to the next element in the sequence. Conversely, you can't coerce the deserializer to operate against string[] because it can't be a SequenceOf and GeneralizedString/IA5String at the same time. +
Type: Upgrade version
+Release Date: 2018-05-08
+Fix Resolution: Microsoft.AspNetCore.Server.Kestrel.Core - 2.0.3,2.1.0;Microsoft.AspNetCore.Server.Kestrel.Transport.Abstractions - 2.0.3,2.1.0;Microsoft.AspNetCore.Server.Kestrel.Transport.Libuv - 2.0.3,2.1.0;Microsoft.AspNetCore.All - 2.0.8,2.1.0
+ + -### Sample +
-var req = AsnSerializer.Deserialize
ASP.NET Core components for working with the IIS AspNetCoreModule.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.iisintegration/2.0.0/microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg
+
-[StructLayout(LayoutKind.Sequential)]
-public struct PaData
-{
- // padata-type[1] Int32,
- // padata-value[2] OCTET STRING -- might be encoded AP-REQ
+Dependency Hierarchy:
+ - dotnetcore-sample-1.0.0 (Root Library)
+ - microsoft.aspnetcore.2.0.0.nupkg
+ - :x: **microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg** (Vulnerable Library)
- [ExpectedTag(1, ExplicitTag = true)]
- public int paDataType;
- [ExpectedTag(2, ExplicitTag = true), OctetString]
- public ReadOnlyMemory
ASP.NET Core hosting infrastructure and startup logic for web applications.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.hosting.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.hosting/2.0.0/microsoft.aspnetcore.hosting.2.0.0.nupkg
+-[StructLayout(LayoutKind.Sequential)] -public struct PrincipalName -{ - // name-type[0] Int32, - // name-string[1] SEQUENCE OF KerberosString +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.hosting.2.0.0.nupkg** (Vulnerable Library) - [ExpectedTag(0, ExplicitTag = true)] - public int nameType; - [ExpectedTag(1)] - public KerberosString[] nameString; +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ - // this does not work either - // [ExpectedTag(1), GeneralizedString] - // public string[] nameString; -} + -[StructLayout(LayoutKind.Sequential)] -public struct KerberosString -{ - [GeneralizedString] - public string name; -} +### Vulnerability Details ++ +ASP.NET Core 1.0. 1.1, and 2.0 allow an elevation of privilege vulnerability due to how ASP.NET web applications handle web requests, aka ""ASP.NET Core Elevation Of Privilege Vulnerability"". This CVE is unique from CVE-2018-0784. -[StructLayout(LayoutKind.Sequential)] -public struct HostAddress -{ - // addr-type[0] Int32, - // address[1] OCTET STRING +
Publish Date: 2018-03-14 +
URL: CVE-2018-0808
+ - [ExpectedTag(0, ExplicitTag = true)] - public int addrType; + - [ExpectedTag(1, ExplicitTag = true), OctetString] - public ReadOnlyMemory-[StructLayout(LayoutKind.Sequential)] -public struct EncryptedData -{ - // EncryptedData::= SEQUENCE { - // etype[0] Int32 -- EncryptionType --, - // kvno[1] UInt32 OPTIONAL, - // cipher[2] OCTET STRING -- ciphertext - // } +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + - [ExpectedTag(0, ExplicitTag = true)] - public int etype; + - [ExpectedTag(1, ExplicitTag = true), OptionalValue] - public int kvno; +### Suggested Fix +
- [ExpectedTag(2, ExplicitTag = true), OctetString]
- public ReadOnlyMemory Type: Upgrade version Origin: https://portal.msrc.microsoft.com/en-US/security-guidance/advisory/CVE-2018-0808 Release Date: 2018-03-14 Fix Resolution: Microsoft.AspNetCore.Server.IISIntegration - 2.1.0, Microsoft.AspNetCore.Hosting - 2.1.0
JavaScript library for DOM operations
-Library home page: https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.2/jquery.min.js
-Path to vulnerable library: /app/src/main/webapp/resources/js/jquery-1.11.2.min.js
+Core components of ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
Dependency Hierarchy: - - :x: **jquery-1.11.2.min.js** (Vulnerable Library) -
Found in HEAD commit: 919af4911488ae49a0463c8d639240781e4088b0
+ - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) +Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
Found in base branch: main
--jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. +Denial of service vulnerability in ASP.NET Core when a malformed request is terminated. -
Publish Date: 2019-04-20 -
URL: CVE-2019-11358
+Publish Date: 2018-07-10 +
URL: WS-2018-0607
-Base Score Metrics: @@ -3519,55 +7210,66 @@ Base Score Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - - User Interaction: Required - - Scope: Changed + - User Interaction: None + - Scope: Unchanged - Impact Metrics: - - Confidentiality Impact: Low - - Integrity Impact: Low - - Availability Impact: None + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High
For more information on CVSS3 Scores, click here. -
Type: Upgrade version
-Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358
-Release Date: 2019-04-20
-Fix Resolution: 3.4.0
+Release Date: 2018-07-10
+Fix Resolution: Microsoft.AspNetCore.Server.Kestrel.Core - 2.1.2
-JavaScript library for DOM operations
-Library home page: https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.2/jquery.min.js
-Path to vulnerable library: /app/src/main/webapp/resources/js/jquery-1.11.2.min.js
+ASP.NET Core default HTTP feature implementations.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.http.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.http/2.0.0/microsoft.aspnetcore.http.2.0.0.nupkg
Dependency Hierarchy: - - :x: **jquery-1.11.2.min.js** (Vulnerable Library) -
Found in HEAD commit: 919af4911488ae49a0463c8d639240781e4088b0
+ - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - microsoft.aspnetcore.hosting.2.0.0.nupkg + - :x: **microsoft.aspnetcore.http.2.0.0.nupkg** (Vulnerable Library) +Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
Found in base branch: main
--jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. +A security feature bypass vulnerability exists in the way Microsoft ASP.NET Core parses encoded cookie names.The ASP.NET Core cookie parser decodes entire cookie strings which could allow a malicious attacker to set a second cookie with the name being percent encoded.The security update addresses the vulnerability by fixing the way the ASP.NET Core cookie parser handles encoded names., aka 'Microsoft ASP.NET Core Security Feature Bypass Vulnerability'. -
Publish Date: 2019-04-20 -
URL: CVE-2019-11358
+Publish Date: 2020-09-11 +
URL: CVE-2020-1045
-Base Score Metrics: @@ -3575,1436 +7277,7593 @@ Base Score Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - - User Interaction: Required - - Scope: Changed + - User Interaction: None + - Scope: Unchanged - Impact Metrics: - - Confidentiality Impact: Low - - Integrity Impact: Low + - Confidentiality Impact: None + - Integrity Impact: High - Availability Impact: None
For more information on CVSS3 Scores, click here. -
Type: Upgrade version
-Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358
-Release Date: 2019-04-20
-Fix Resolution: 3.4.0
+Release Date: 2020-09-11
+Fix Resolution: Microsoft.AspNetCore.App - 2.1.22, Microsoft.AspNetCore.All - 2.1.22,Microsoft.NETCore.App - 2.1.22, Microsoft.AspNetCore.Http - 2.1.22
-Core components of ASP.NET Core Kestrel cross-platform web server.
+Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+- +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.2.0.0.nupkg + - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg + - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -Zalenium Image Version(s): -""dosel/zalenium:latest"" -If using docker-compose, version: - -OS: - -Docker Command to start Zalenium: - -version: '2' + -services: - - zalenium: - image: ""dosel/zalenium:latest"" - container_name: zalenium - hostname: zalenium - tty: true - volumes: - - /tmp/videos:/home/seluser/videos - - /var/run/docker.sock:/var/run/docker.sock - - /usr/bin/docker:/usr/bin/docker - ports: - - 4444:4444 - extra_hosts: - - ""myapp.mydomain.fr:172.25.0.10"" - command: > - start --desiredContainers 2 - --maxDockerSeleniumContainers 8 - --screenWidth 1680 --screenHeight 1050 - --timeZone ""Europe/Paris"" - --videoRecordingEnabled true - --sauceLabsEnabled false - --browserStackEnabled false - --testingBotEnabled false - --startTunnel false - environment: - - HOST_UID - - HOST_GID - - SAUCE_USERNAME - - SAUCE_ACCESS_KEY - - BROWSER_STACK_USER - - BROWSER_STACK_KEY - - TESTINGBOT_KEY - - TESTINGBOT_SECRET - - LANG=fr_FR.UTF-8 - - LANGUAGE=fr_FR.UTF-8 - - LC_ALL=fr_FR.UTF-8 - networks: - mynet: - ipv4_address: 172.25.0.15 - - -I set : -chromeOptions.addArguments(""lang=fr_FR""); -capability.setCapability(ChromeOptions.CAPABILITY, chromeOptions); -before invoke webdrvier : -new RemoteWebDriver(new URL(this.hubUrl), capability); - -My browser is always in english and not take arguement ""lang=fr"" - -Thank's for you help and sorry for by bad english :) ",1.0,"Can't choose chrome's language - **Please make sure that you provide enough information for us to help you with this issue. Thank you!** - - - -Zalenium Image Version(s): -""dosel/zalenium:latest"" -If using docker-compose, version: - -OS: - -Docker Command to start Zalenium: - -version: '2' - -services: +### Vulnerability Details +- zalenium: - image: ""dosel/zalenium:latest"" - container_name: zalenium - hostname: zalenium - tty: true - volumes: - - /tmp/videos:/home/seluser/videos - - /var/run/docker.sock:/var/run/docker.sock - - /usr/bin/docker:/usr/bin/docker - ports: - - 4444:4444 - extra_hosts: - - ""myapp.mydomain.fr:172.25.0.10"" - command: > - start --desiredContainers 2 - --maxDockerSeleniumContainers 8 - --screenWidth 1680 --screenHeight 1050 - --timeZone ""Europe/Paris"" - --videoRecordingEnabled true - --sauceLabsEnabled false - --browserStackEnabled false - --testingBotEnabled false - --startTunnel false - environment: - - HOST_UID - - HOST_GID - - SAUCE_USERNAME - - SAUCE_ACCESS_KEY - - BROWSER_STACK_USER - - BROWSER_STACK_KEY - - TESTINGBOT_KEY - - TESTINGBOT_SECRET - - LANG=fr_FR.UTF-8 - - LANGUAGE=fr_FR.UTF-8 - - LC_ALL=fr_FR.UTF-8 - networks: - mynet: - ipv4_address: 172.25.0.15 - - -I set : -chromeOptions.addArguments(""lang=fr_FR""); -capability.setCapability(ChromeOptions.CAPABILITY, chromeOptions); -before invoke webdrvier : -new RemoteWebDriver(new URL(this.hubUrl), capability); - -My browser is always in english and not take arguement ""lang=fr"" - -Thank's for you help and sorry for by bad english :) ",0,can t choose chrome s language please make sure that you provide enough information for us to help you with this issue thank you zalenium image version s dosel zalenium latest if using docker compose version os docker command to start zalenium version services zalenium image dosel zalenium latest container name zalenium hostname zalenium tty true volumes tmp videos home seluser videos var run docker sock var run docker sock usr bin docker usr bin docker ports extra hosts myapp mydomain fr command start desiredcontainers maxdockerseleniumcontainers screenwidth screenheight timezone europe paris videorecordingenabled true saucelabsenabled false browserstackenabled false testingbotenabled false starttunnel false environment host uid host gid sauce username sauce access key browser stack user browser stack key testingbot key testingbot secret lang fr fr utf language fr fr utf lc all fr fr utf networks mynet address i set chromeoptions addarguments lang fr fr capability setcapability chromeoptions capability chromeoptions before invoke webdrvier new remotewebdriver new url this huburl capability my browser is always in english and not take arguement lang fr thank s for you help and sorry for by bad english ,0 -12228,14442108234.0,IssuesEvent,2020-12-07 17:41:07,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,parser: STRAIGHT_JOIN and SQL_SMALL_RESULT cannot be used at the same time,component/parser severity/moderate type/bug type/compatibility type/duplicate,"## Bug Report +ASP.NET Core and Visual Studio Denial of Service Vulnerability -Please answer these questions before submitting your issue. Thanks! +
Publish Date: 2021-01-12 +
URL: CVE-2021-1723
+ -1. What did you do? -``` -create table t; -insert into t values(1); -``` -2. What did you expect to see? -Results from Mysql: -``` -mysql> select STRAIGHT_JOIN SQL_SMALL_RESULT count(*) from t; -+----------+ -| count(*) | -+----------+ -| 1 | -+----------+ -1 row in set (0.00 sec) -``` + -3. What did you see instead? -Results: from TiDB: -``` -select STRAIGHT_JOIN SQL_SMALL_RESULT count(*) from t; -ERROR 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your TiDB version for the right syntax to use line 1 column 37 near ""SQL_SMALL_RESULT count(*) from t"" -``` +### CVSS 3 Score Details (7.5) +-4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? -[tidb@VM2 bin]$ ./tidb-server -V -2019-11-26 18:27:18.059136 I | maxprocs: Leaving GOMAXPROCS=8: CPU quota undefined -Release Version: -Git Commit Hash: -Git Branch: -UTC Build Time: 2019-11-22 02:54:53 -GoVersion: go version go1.13.3 linux/amd64 -Race Enabled: false -TiKV Min Version: v3.0.0-60965b006877ca7234adaced7890d7b029ed1306 -Check Table Before Drop: false -",True,"parser: STRAIGHT_JOIN and SQL_SMALL_RESULT cannot be used at the same time - ## Bug Report - -Please answer these questions before submitting your issue. Thanks! - -1. What did you do? -``` -create table t; -insert into t values(1); -``` -2. What did you expect to see? -Results from Mysql: -``` -mysql> select STRAIGHT_JOIN SQL_SMALL_RESULT count(*) from t; -+----------+ -| count(*) | -+----------+ -| 1 | -+----------+ -1 row in set (0.00 sec) -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + -3. What did you see instead? -Results: from TiDB: -``` -select STRAIGHT_JOIN SQL_SMALL_RESULT count(*) from t; -ERROR 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your TiDB version for the right syntax to use line 1 column 37 near ""SQL_SMALL_RESULT count(*) from t"" -``` + -4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? -[tidb@VM2 bin]$ ./tidb-server -V -2019-11-26 18:27:18.059136 I | maxprocs: Leaving GOMAXPROCS=8: CPU quota undefined -Release Version: -Git Commit Hash: -Git Branch: -UTC Build Time: 2019-11-22 02:54:53 -GoVersion: go version go1.13.3 linux/amd64 -Race Enabled: false -TiKV Min Version: v3.0.0-60965b006877ca7234adaced7890d7b029ed1306 -Check Table Before Drop: false -",1,parser straight join and sql small result cannot be used at the same time bug report please answer these questions before submitting your issue thanks what did you do create table t insert into t values what did you expect to see results from mysql mysql select straight join sql small result count from t count row in set sec what did you see instead results from tidb select straight join sql small result count from t error you have an error in your sql syntax check the manual that corresponds to your tidb version for the right syntax to use line column near sql small result count from t what version of tidb are you using tidb server v or run select tidb version on tidb tidb server v i maxprocs leaving gomaxprocs cpu quota undefined release version git commit hash git branch utc build time goversion go version linux race enabled false tikv min version check table before drop false ,1 -136962,20023377990.0,IssuesEvent,2022-02-01 18:32:44,psf/black,https://api.github.com/repos/psf/black,reopened,Line splitting on conditional operators should take precendence over splitting on function parameters,R: duplicate T: design F: linebreak,"Black should split conditionals at conditional operators (i.e. `and`, `or`, etc.) when possible, instead of splitting function parameters over multiple lines. +### Suggested Fix +-Currently, this is how black formats a long conditional: -``` -if str( - self.user.id -) != self.target_user_id and not CheckUserManagementViewPermission.user_has_permission( - self.user -): - pass -``` +
Type: Upgrade version
+Origin: https://nvd.nist.gov/vuln/detail/CVE-2021-1723
+Release Date: 2021-01-12
+Fix Resolution: Microsoft.AspNetCore.App.Runtime.win-arm64 - 5.0.2;LiveReloadServer - 1.1.0;Plugga.Core - 1.0.2;Maple.Branch.Module - 1.0.4;Microsoft.AspNetCore.Components.WebAssembly.Server - 5.0.1,5.0.0-rc.1.20451.17;AspNetCoreRuntime.5.0.x64 - 5.0.2;AspNetCoreRuntime.5.0.x86 - 5.0.2;Microsoft.AspNetCore.App.Runtime.osx-x64 - 5.0.2,3.1.10;GrazeDocs - 2.0.1;Microsoft.AspNetCore.App.Runtime.linux-musl-arm - 5.0.2;Microsoft.AspNetCore.App.Runtime.linux-musl-x64 - 5.0.2,3.1.10;YHWins.Template - 1.1.0;Microsoft.AspNetCore.App.Runtime.linux-musl-arm64 - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Runtime.linux-arm64 - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Ref - 3.1.10,6.0.0-rc.1.21452.15;Microsoft.AspNetCore.Blazor.DevServer - 3.2.0-preview1.20073.1,3.1.0-preview4.19579.2;Microsoft.AspNetCore.App.Runtime.linux-arm - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Runtime.linux-x64 - 3.1.10,5.0.2;stankins.console - 2020.12.20-beta298;Toolbelt.Blazor.DevServer.WithCssLiveReloader - 5.0.1,5.0.0-rc.1.20451.17;DragonFire.Server - 0.0.1-alpha.0;PoExtractor.OrchardCore - 0.5.0-rc2-16220;Microsoft.AspNetCore.App.Runtime.win-arm - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Runtime.win-x64 - 3.1.10,5.0.2;Microsoft.AspNetCore.App.Runtime.win-x86 - 3.1.10,5.0.2;HuLu.Template.Api - 1.0.2;AspNetCoreRuntime.3.1.x64 - 3.1.10;AspNetCoreRuntime.3.1.x86 - 3.1.10;Microsoft.AspNetCore.Components.WebAssembly.DevServer - 5.0.0-rc.1.20451.17,5.0.1;Microsoft.AspNetCore.App.Runtime.win-arm64 - 3.1.10;lingman-webapi - 0.0.18
-Ideally, it should be formatted like this: -``` -if ( - str(self.user.id) != self.target_user_id - and not CheckUserManagementViewPermission.user_has_permission( - self.user - ) -): - pass -``` + -This appears to be a general issue with splitting conditionals and it looks like it's related to these issues: https://github.com/psf/black/issues/815 https://github.com/psf/black/issues/1050",1.0,"Line splitting on conditional operators should take precendence over splitting on function parameters - Black should split conditionals at conditional operators (i.e. `and`, `or`, etc.) when possible, instead of splitting function parameters over multiple lines. + -Currently, this is how black formats a long conditional: -``` -if str( - self.user.id -) != self.target_user_id and not CheckUserManagementViewPermission.user_has_permission( - self.user -): - pass -``` +Provides classes to support the creation and validation of XML digital signatures. The classes in th...
+Library home page: https://api.nuget.org/packages/system.security.cryptography.xml.4.4.0.nupkg
+Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
+Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.4.0/system.security.cryptography.xml.4.4.0.nupkg
+-1. Add a new quiz with at least one question -2. Preview quiz all the way through, returning to module overview page -3. Do something else: E.g., go to edit quiz page, or do manual refresh of module overview page, then return to module overview page +Dependency Hierarchy: + - dotnetcore-sample-1.0.0 (Root Library) + - microsoft.aspnetcore.mvc.2.0.0.nupkg + - microsoft.aspnetcore.mvc.viewfeatures.2.0.0.nupkg + - microsoft.aspnetcore.antiforgery.2.0.0.nupkg + - microsoft.aspnetcore.dataprotection.2.0.0.nupkg + - :x: **system.security.cryptography.xml.4.4.0.nupkg** (Vulnerable Library) +
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
+Found in base branch: main
+ -The previewed quiz is now shown as having been run, and cannot be run again.",0,previewed quiz set to already run to replicate add a new quiz with at least one question preview quiz all the way through returning to module overview page do something else e g go to edit quiz page or do manual refresh of module overview page then return to module overview page the previewed quiz is now shown as having been run and cannot be run again ,0 -11861,13995564192.0,IssuesEvent,2020-10-28 03:36:04,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,Support multiple table rename (swap table use case),challenge-program difficulty/medium high-performance picked sig/DDL type/compatibility type/enhancement,"## Feature Request + -**Is your feature request related to a problem? Please describe:** +### Vulnerability Details ++ +Microsoft is releasing this security advisory to provide information about a vulnerability in .NET Core 3.1 and .NET 6.0. An information disclosure vulnerability exists in .NET Core 3.1 and .NET 6.0 that could lead to unauthorized access of privileged information. -I know that multiple DDL changes are currently on the roadmap. My feature request is to support a much smaller subset, which is an atomic rename of two tables. +## Affected software -This supports the use case of swapping tables, something that occurs frequently in the replication stream due to its usage by tools such as `pt-online-schema-change`. +* Any .NET 6.0 application running on .NET 6.0.7 or earlier. +* Any .NET Core 3.1 applicaiton running on .NET Core 3.1.27 or earlier. -**Describe the feature you'd like:** +## Patches -``` -mysql> CREATE TABLE t1 (a int); -Query OK, 0 rows affected (0.01 sec) +* If you're using .NET 6.0, you should download and install Runtime 6.0.8 or SDK 6.0.108 (for Visual Studio 2022 v17.1) from https://dotnet.microsoft.com/download/dotnet-core/6.0. +* If you're using .NET Core 3.1, you should download and install Runtime 3.1.28 (for Visual Studio 2019 v16.9) from https://dotnet.microsoft.com/download/dotnet-core/3.1. -mysql> CREATE TABLE t2 (a int, b int); -Query OK, 0 rows affected (0.01 sec) +
Publish Date: 2022-08-09 +
URL: CVE-2022-34716
+ -mysql> RENAME TABLE t1 to t1_old, t2 to t1; <--- should work -ERROR 1105 (HY000): can't run multi schema change -``` + -**Describe alternatives you've considered:** +### CVSS 3 Score Details (5.9) +-The alternative is to do this non-atomically (breaks apps) or under a global lock (not yet supported by TiDB). So there is not really a good alternative. +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: High + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: None + - Availability Impact: None +
+For more information on CVSS3 Scores, click here. + -**Teachability, Documentation, Adoption, Migration Strategy:** + -MySQL compatible behavior. +### Suggested Fix +-## Score -1500 +
Type: Upgrade version
+Origin: https://github.com/advisories/GHSA-2m65-m22p-9wjw
+Release Date: 2022-08-09
+Fix Resolution: Microsoft.AspNetCore.App.Runtime.linux-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.osx-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x86 - 3.1.28,6.0.8;System.Security.Cryptography.Xml - 4.7.1,6.0.1
-## SIG Slack Channel -You can join [#sig-ddl ](https://app.slack.com/client/TH91JCS4W/CMAKWBNJU)on slack in your spare time to discuss and get help with mentors or others. + -## Mentor(s) -* @zimulala + -## Recommended Skills -* DDL -* Golang +General data-binding functionality for Jackson: works on core streaming API
+Library home page: http://github.com/FasterXML/jackson
+Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml
+Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar
+-3. What did you see instead? -TiDB -``` -mysql> SELECT PERIOD_diff(-00013,1); -+-----------------------+ -| PERIOD_diff(-00013,1) | -+-----------------------+ -| 2213609288845122194 | -+-----------------------+ -1 row in set (0.00 sec) -``` +Dependency Hierarchy: + - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) +
Found in HEAD commit: e2b236143990842a0d83d97532011829192916a7
+Found in base branch: master
+ ++ +A flaw was discovered in FasterXML jackson-databind in all versions before 2.9.10 and 2.10.0, where it would permit polymorphic deserialization of malicious objects using the xalan JNDI gadget when used in conjunction with polymorphic type handling methods such as `enableDefaultTyping()` or when @JsonTypeInfo is using `Id.CLASS` or `Id.MINIMAL_CLASS` or in any other way which ObjectMapper.readValue might instantiate objects from unsafe sources. An attacker could use this flaw to execute arbitrary code. +
Publish Date: 2020-03-02 +
URL: CVE-2019-14893
+ +-4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? -``` -mysql> select tidb_version(); -+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| tidb_version() | -+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| Release Version: v3.0.0-beta.1-218-g2f7241138-dirty -Git Commit Hash: 2f7241138adfc434f4b199ecae68fb83ad33b36b -Git Branch: myfeature-perioddiff -UTC Build Time: 2019-05-07 03:36:21 -GoVersion: go version go1.11.4 darwin/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -Check Table Before Drop: false | -+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -1 row in set (0.00 sec) -``` -",1,function period diff is not compatible with mysql bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error function period diff is not compatible with mysql what did you expect to see mysql mysql select period diff error incorrect arguments to period diff what did you see instead tidb mysql select period diff period diff row in set sec what version of tidb are you using tidb server v or run select tidb version on tidb mysql select tidb version tidb version release version beta dirty git commit hash git branch myfeature perioddiff utc build time goversion go version darwin race enabled false tikv min version alpha check table before drop false row in set sec ,1 -301115,9216196333.0,IssuesEvent,2019-03-11 07:11:51,etternagame/etterna,https://api.github.com/repos/etternagame/etterna,closed,Misc Bugs here and there.,Priority: Low,"I made a theme that uses etterna's fallback theme. I went though trying to debug a few things here and there. I found a few issues. +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-(Missing Metrics) +
Type: Upgrade version
+Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14893
+Release Date: 2020-03-02
+Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0
-ScreenSelectPlayMode::ChoiceNonstop="""" -ScreenSelectPlayMode::ChoiceOni="""" -ScreenSelectPlayMode::ChoiceEndless="""" -OptionExplanations::FullTapExplosions=""""General data-binding functionality for Jackson: works on core streaming API
+Library home page: http://github.com/FasterXML/jackson
+Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml
+Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar
+-(Crash) +Dependency Hierarchy: + - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) +
Found in HEAD commit: e2b236143990842a0d83d97532011829192916a7
+Found in base branch: master
+ ++ +A flaw was discovered in FasterXML jackson-databind in all versions before 2.9.10 and 2.10.0, where it would permit polymorphic deserialization of malicious objects using the xalan JNDI gadget when used in conjunction with polymorphic type handling methods such as `enableDefaultTyping()` or when @JsonTypeInfo is using `Id.CLASS` or `Id.MINIMAL_CLASS` or in any other way which ObjectMapper.readValue might instantiate objects from unsafe sources. An attacker could use this flaw to execute arbitrary code. -Couples Mode crashes the game. Needs to be removed. +
Publish Date: 2020-03-02 +
URL: CVE-2019-14893
+ +-(Concerns) +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +
-Verses and Doubles work but should be removed. (Since player 2 was removed, likely, this will not work in doubles/verses.)
-Nonstop, Oni, and Endless should be removed. (If done, you can ignore the metrics then.)
-
------------------------------------------------------------------------------------------------------------
+
Type: Upgrade version
+Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14893
+Release Date: 2020-03-02
+Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0
-More bugs found. The following seemed to have been removed from the game: + +Tiny ms conversion utility
+Library home page: https://registry.npmjs.org/ms/-/ms-0.7.1.tgz
+Path to dependency file: brackets/package.json
+Path to vulnerable library: brackets/node_modules/tiny-lr/node_modules/ms/package.json,brackets/node_modules/body-parser/node_modules/ms/package.json
+-Couples Mode crashes the game. Needs to be removed. +Dependency Hierarchy: + - grunt-contrib-watch-1.0.0.tgz (Root Library) + - tiny-lr-0.2.1.tgz + - debug-2.2.0.tgz + - :x: **ms-0.7.1.tgz** (Vulnerable Library) +
Found in HEAD commit: 486c979c9181a42d1ab9d6be10b160fb1973f21b
+Found in base branch: checkTravis
+ ++ +Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS). -(Concerns) +
Publish Date: 2017-04-12 +
URL: WS-2017-0247
+ +
-Verses and Doubles work but should be removed. (Since player 2 was removed, likely, this will not work in doubles/verses.)
-Nonstop, Oni, and Endless should be removed. (If done, you can ignore the metrics then.)
+Base Score Metrics:
+- Exploitability Metrics:
+ - Attack Vector: Network
+ - Attack Complexity: Low
+ - Privileges Required: None
+ - User Interaction: None
+ - Scope: Unchanged
+- Impact Metrics:
+ - Confidentiality Impact: None
+ - Integrity Impact: None
+ - Availability Impact: High
+
------------------------------------------------------------------------------------------------------------ +
Type: Upgrade version
+Origin: https://github.com/vercel/ms/pull/89
+Release Date: 2017-04-12
+Fix Resolution: 2.1.1
-More bugs found. The following seemed to have been removed from the game: + +Tiny ms conversion utility
+Library home page: https://registry.npmjs.org/ms/-/ms-0.7.1.tgz
+Path to dependency file: brackets/package.json
+Path to vulnerable library: brackets/node_modules/tiny-lr/node_modules/ms/package.json,brackets/node_modules/body-parser/node_modules/ms/package.json
+-Themes/Til Death/BGAnimations/ScreenSelectMusic decorations/other.lua -src/SongManager.cpp -Docs/Luadoc/Lua.xml -Docs/Luadoc/LuaDocumentation.xml +Dependency Hierarchy: + - grunt-contrib-watch-1.0.0.tgz (Root Library) + - tiny-lr-0.2.1.tgz + - debug-2.2.0.tgz + - :x: **ms-0.7.1.tgz** (Vulnerable Library) +
Found in HEAD commit: 486c979c9181a42d1ab9d6be10b160fb1973f21b
+Found in base branch: checkTravis
+ ++ +Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS). -SONGMAN:GetNumAdditionalCourses(); +
Publish Date: 2017-04-12 +
URL: WS-2017-0247
+ +-Docs/Luadoc/LuaDocumentation.xml +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-SONGMAN:GetNumLockedSongs(); +
Type: Upgrade version
+Origin: https://github.com/vercel/ms/pull/89
+Release Date: 2017-04-12
+Fix Resolution: 2.1.1
-Docs/Changelog_sm5.txt -Docs/Luadoc/LuaDocumentation.xml + +Simple GIT interface for node.js
+Library home page: https://registry.npmjs.org/simple-git/-/simple-git-1.131.0.tgz
+Path to dependency file: /package.json
+Path to vulnerable library: /node_modules/simple-git/package.json
+-Update: Hmm... Seems GAMESTATE:GetCurrentStage(); was removed as well? Oh well. +Dependency Hierarchy: + - :x: **simple-git-1.131.0.tgz** (Vulnerable Library) +
Found in HEAD commit: 4fc776e2b95c8b497f6994cb2165365562ae1f82
+Found in base branch: master
+ ++ +The package simple-git before 3.15.0 are vulnerable to Remote Code Execution (RCE) when enabling the ext transport protocol, which makes it exploitable via clone() method. This vulnerability exists due to an incomplete fix of [CVE-2022-24066](https://security.snyk.io/vuln/SNYK-JS-SIMPLEGIT-2434306). -Docs/Changelog_sm5.txt -Docs/Luadoc/LuaDocumentation.xml +
Publish Date: 2022-12-06 +
URL: CVE-2022-25912
+ +-Update: I am unsure but, I think this is removed. PREFSMAN:GetPreference(""PercentageScoring""); Although, a few references to it. +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-Themes/_fallback/Languages/zh.ini -Themes/_fallback/Languages/es.ini -Themes/_fallback/Languages/fr.ini -Themes/_fallback/Languages/ja.ini -Themes/_fallback/Languages/nl.ini -Themes/_fallback/Languages/pl.ini -Themes/_fallback/Languages/de.ini -Themes/_fallback/Languages/en.ini -Docs/old_changelog.txt -Docs/Changelog_sm-ssc.txt -Themes/_fallback/metrics.ini +
Type: Upgrade version
+Origin: https://www.cve.org/CVERecord?id=CVE-2022-25912
+Release Date: 2022-12-06
+Fix Resolution: 3.15.0
-Update: Unsure of this but likely my theme. Check into metric ComboContinuesBetweenSongs. + +Simple GIT interface for node.js
+Library home page: https://registry.npmjs.org/simple-git/-/simple-git-1.131.0.tgz
+Path to dependency file: /package.json
+Path to vulnerable library: /node_modules/simple-git/package.json
+-1. What did you do? +Dependency Hierarchy: + - :x: **simple-git-1.131.0.tgz** (Vulnerable Library) +
Found in HEAD commit: 4fc776e2b95c8b497f6994cb2165365562ae1f82
+Found in base branch: master
+ ++ +The package simple-git before 3.15.0 are vulnerable to Remote Code Execution (RCE) when enabling the ext transport protocol, which makes it exploitable via clone() method. This vulnerability exists due to an incomplete fix of [CVE-2022-24066](https://security.snyk.io/vuln/SNYK-JS-SIMPLEGIT-2434306). -``` -CREATE TABLE t1 (id int(11) NOT NULL AUTO_INCREMENT, PRIMARY KEY (`id`)); -INSERT INTO t1(`id`) VALUES(3); -CREATE TABLE t2 (title text); -INSERT INTO t2(`title`) VALUES(""3.0.1""); -SELECT * FROM t1 JOIN t2 ON t1.id=t2.title; -``` - -2. What did you expect to see? +
Publish Date: 2022-12-06 +
URL: CVE-2022-25912
+ +-* MySQL 8.0 +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-``` -mysql> CREATE TABLE t1 (id int(11) NOT NULL AUTO_INCREMENT, PRIMARY KEY (`id`)); -Query OK, 0 rows affected, 1 warning (0.01 sec) +
Type: Upgrade version
+Origin: https://www.cve.org/CVERecord?id=CVE-2022-25912
+Release Date: 2022-12-06
+Fix Resolution: 3.15.0
-mysql> INSERT INTO t1(`id`) VALUES(3); -Query OK, 1 row affected (0.00 sec) + +a glob matcher in javascript
+Library home page: https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz
+Path to dependency file: /package.json
+Path to vulnerable library: /node_modules/recursive-readdir/node_modules/minimatch/package.json
+-The fix is to change `object` to `PyObject*` for both args and kwargs. +Dependency Hierarchy: + - react-scripts-5.0.1.tgz (Root Library) + - react-dev-utils-12.0.1.tgz + - recursive-readdir-2.2.2.tgz + - :x: **minimatch-3.0.4.tgz** (Vulnerable Library) +
Found in base branch: master
+ ++ +A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service. -Please see below for details: +
Publish Date: 2022-10-17 +
URL: CVE-2022-3517
+ +-```cython -# cython: language_level=3 -from cpython cimport newfunc, type as cpytype, Py_TYPE +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-cdef class X: - cdef int i - def __init__(self, i): - self.i = i - def __repr__(self): - return 'X(%d)' % self.i +
Type: Upgrade version
+Release Date: 2022-10-17
+Fix Resolution: minimatch - 3.0.5
-cdef newfunc _orig_tp_new = Py_TYPE(X(0)).tp_new + +a glob matcher in javascript
+Library home page: https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz
+Path to dependency file: /package.json
+Path to vulnerable library: /node_modules/recursive-readdir/node_modules/minimatch/package.json
++Dependency Hierarchy: + - react-scripts-5.0.1.tgz (Root Library) + - react-dev-utils-12.0.1.tgz + - recursive-readdir-2.2.2.tgz + - :x: **minimatch-3.0.4.tgz** (Vulnerable Library) +
Found in base branch: master
+ ++ +A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service. -x = X(123) -print(x) -``` +
Publish Date: 2022-10-17 +
URL: CVE-2022-3517
+ +-```console -(neo) (py3.venv) (g.env) kirr@deca:~/src/tools/go/pygolang$ cythonize -i x.pyx -Compiling /home/kirr/src/tools/go/pygolang/x.pyx because it changed. -[1/1] Cythonizing /home/kirr/src/tools/go/pygolang/x.pyx -running build_ext -building 'x' extension -... -x86_64-linux-gnu-gcc -pthread -Wno-unused-result -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -ffile-prefix-map=/build/python3.9-RNBry6/python3.9-3.9.2=. -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -g -ffile-prefix-map=/build/python3.9-RNBry6/python3.9-3.9.2=. -fstack-protector-strong -Wformat -Werror=format-security -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -I/home/kirr/src/wendelin/venv/py3.venv/include -I/usr/include/python3.9 -c /home/kirr/src/tools/go/pygolang/x.c -o /home/kirr/src/tools/go/pygolang/tmpqkz1r96s/home/kirr/src/tools/go/pygolang/x.o -x86_64-linux-gnu-gcc -pthread -shared -Wl,-O1 -Wl,-Bsymbolic-functions -Wl,-z,relro -g -fwrapv -O2 -Wl,-z,relro -g -fwrapv -O2 -g -ffile-prefix-map=/build/python3.9-RNBry6/python3.9-3.9.2=. -fstack-protector-strong -Wformat -Werror=format-security -Wdate-time -D_FORTIFY_SOURCE=2 /home/kirr/src/tools/go/pygolang/tmpqkz1r96s/home/kirr/src/tools/go/pygolang/x.o -o /home/kirr/src/tools/go/pygolang/x.cpython-39-x86_64-linux-gnu.so -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +
-```console
-(neo) (py3.venv) (g.env) kirr@deca:~/src/tools/go/pygolang$ python -c 'import x'
-Ошибка сегментирования (стек памяти сброшен на диск)
-```
-
-```console
-(neo) (py3.venv) (g.env) kirr@deca:~/src/tools/go/pygolang$ gdb python core
-...
-Reading symbols from python...
-Reading symbols from /usr/lib/debug/.build-id/f9/02f8a561c3abdb9c8d8c859d4243bd8c3f928f.debug...
-[New LWP 218557]
-[Thread debugging using libthread_db enabled]
-Using host libthread_db library ""/lib/x86_64-linux-gnu/libthread_db.so.1"".
-Core was generated by `python -c import x'.
-Program terminated with signal SIGSEGV, Segmentation fault.
-#0 _Py_INCREF (op=0x0) at /usr/include/python3.9/object.h:408
-408 op->ob_refcnt++;
-
-(gdb) bt 5
-#0 _Py_INCREF (op=0x0) at /usr/include/python3.9/object.h:408
-#1 __pyx_f_1x__trace_tp_new (__pyx_v_cls=0x7f5ce75e6880 <__pyx_type_1x_X>, __pyx_v_args=(123,), __pyx_v_kw=0x0) at /home/kirr/src/tools/go/pygolang/x.c:1986
-#2 0x000000000051dd7e in type_call (type=type@entry=0x7f5ce75e6880 <__pyx_type_1x_X>, args=args@entry=(123,), kwds=kwds@entry=0x0)
- at ../Objects/typeobject.c:1014
-#3 0x00007f5ce75df8d4 in __Pyx_PyObject_Call (func= Type: Upgrade version Release Date: 2022-10-17 Fix Resolution: minimatch - 3.0.5
+jackson-databind-2.9.5.jar
General data-binding functionality for Jackson: works on core streaming API
+Library home page: http://github.com/FasterXML/jackson
+Path to dependency file: Teste/liferay-portal/modules/etl/talend/talend-runtime/pom.xml
+Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar
-
Julia Cartwright's fork of linux-stable-rt.git
-Library home page: https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git
-Found in HEAD commit: 9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69
- -Found in base branch: master
+Dependency Hierarchy: + - components-api-0.25.3.jar (Root Library) + - daikon-0.27.0.jar + - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)General data-binding functionality for Jackson: works on core streaming API
+Library home page: http://github.com/FasterXML/jackson
+Path to dependency file: Teste/opentelemetry-java/sdk_contrib/aws_v1_support/build.gradle
+Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.7.3/c8bad9f57a5d298280f8bc0efcb7bf8393a41534/jackson-databind-2.6.7.3.jar
- /drivers/staging/rtl8188eu/os_dep/ioctl_linux.c
-
/drivers/staging/rtl8188eu/os_dep/ioctl_linux.c
-
Found in HEAD commit: 64dde89c50c07496423c4d4a865f2e16b92399ad
-rtw_wx_set_scan in drivers/staging/rtl8188eu/os_dep/ioctl_linux.c in the Linux kernel through 5.11.6 allows writing beyond the end of the ->ssid[] array. NOTE: from the perspective of kernel.org releases, CVE IDs are not normally used for drivers/staging/* (unfinished work); however, system integrators may have situations in which a drivers/staging issue is relevant to their own customer base. +A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload. -
Publish Date: 2021-03-17 -
URL: CVE-2021-28660
+Publish Date: 2019-10-12 +
URL: CVE-2019-17531
Base Score Metrics:
- Exploitability Metrics:
- - Attack Vector: Local
+ - Attack Vector: Network
- Attack Complexity: Low
- - Privileges Required: Low
+ - Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
@@ -5020,56 +14879,68 @@ For more information on CVSS3 Scores, click
Type: Upgrade version Origin: https://www.linuxkernelcves.com/cves/CVE-2021-28660 Release Date: 2021-03-17 Fix Resolution: v5.11.6 Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531 Release Date: 2019-10-12 Fix Resolution: 2.10
-
Julia Cartwright's fork of linux-stable-rt.git
-Library home page: https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git
-Found in HEAD commit: 9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69
+Found in base branch: master
+General data-binding functionality for Jackson: works on core streaming API
+Library home page: http://github.com/FasterXML/jackson
+Path to dependency file: Teste/liferay-portal/modules/etl/talend/talend-runtime/pom.xml
+Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar
++ +Dependency Hierarchy: + - components-api-0.25.3.jar (Root Library) + - daikon-0.27.0.jar + - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
General data-binding functionality for Jackson: works on core streaming API
+Library home page: http://github.com/FasterXML/jackson
+Path to dependency file: Teste/opentelemetry-java/sdk_contrib/aws_v1_support/build.gradle
+Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.7.3/c8bad9f57a5d298280f8bc0efcb7bf8393a41534/jackson-databind-2.6.7.3.jar
- /drivers/staging/rtl8188eu/os_dep/ioctl_linux.c
-
/drivers/staging/rtl8188eu/os_dep/ioctl_linux.c
-
Found in HEAD commit: 64dde89c50c07496423c4d4a865f2e16b92399ad
-rtw_wx_set_scan in drivers/staging/rtl8188eu/os_dep/ioctl_linux.c in the Linux kernel through 5.11.6 allows writing beyond the end of the ->ssid[] array. NOTE: from the perspective of kernel.org releases, CVE IDs are not normally used for drivers/staging/* (unfinished work); however, system integrators may have situations in which a drivers/staging issue is relevant to their own customer base. +A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload. -
Publish Date: 2021-03-17 -
URL: CVE-2021-28660
+Publish Date: 2019-10-12 +
URL: CVE-2019-17531
Base Score Metrics:
- Exploitability Metrics:
- - Attack Vector: Local
+ - Attack Vector: Network
- Attack Complexity: Low
- - Privileges Required: Low
+ - Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
@@ -5085,1543 +14956,9191 @@ For more information on CVSS3 Scores, click
Type: Upgrade version Origin: https://www.linuxkernelcves.com/cves/CVE-2021-28660 Release Date: 2021-03-17 Fix Resolution: v5.11.6 Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531 Release Date: 2019-10-12 Fix Resolution: 2.10
+ +Artifacts: [/tpccbench/nodes=9/cpu=4/chaos/partition](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657161&tab=artifacts#/tpccbench/nodes=9/cpu=4/chaos/partition) +[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpccbench%2Fnodes%3D9%2Fcpu%3D4%2Fchaos%2Fpartition.%2A&sort=title&restgroup=false&display=lastcommented+project) +powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
-4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? -``` -mysql> select tidb_version(); -+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| tidb_version() | -+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -| Release Version: v3.0.0-beta-231-g20463d6-dirty -Git Commit Hash: 20463d6da90fdf12c0d9d18c15dc33a78334882d -Git Branch: master -UTC Build Time: 2019-03-20 03:32:11 -GoVersion: go version go1.12 linux/amd64 -Race Enabled: false -TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e -Check Table Before Drop: false | -+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ -1 row in set (0.00 sec) +Artifacts: [/tpccbench/nodes=9/cpu=4/chaos/partition](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657161&tab=artifacts#/tpccbench/nodes=9/cpu=4/chaos/partition) + +[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpccbench%2Fnodes%3D9%2Fcpu%3D4%2Fchaos%2Fpartition.%2A&sort=title&restgroup=false&display=lastcommented+project) +powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
-See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) -
--[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLogic_cluster_settings.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) -
-See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) -
--[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLogic_cluster_settings.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) -
-
Julia Cartwright's fork of linux-stable-rt.git
-Library home page: https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git
-Found in HEAD commit: 5e23b7f9d2dd0154edd54986754eecd5b5308571
+Artifacts: [/transfer-leases/quit](https://teamcity.cockroachdb.com/viewLog.html?buildId=1864738&tab=artifacts#/transfer-leases/quit) +Related: +- #47259 roachtest: transfer-leases/quit failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) -Found in base branch: master
-A utility library delivering consistency, customization, performance, and extras.
+Library home page: https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz
+Path to dependency file: IoT-Smart-Gateway/web/package.json
+Path to vulnerable library: IoT-Smart-Gateway/web/node_modules/globule/node_modules/lodash/package.json
- /drivers/net/wireless/ath/ath9k/xmit.c
-
Found in HEAD commit: 085c2936b8c1a6952ca6f0db9eaf75ff33e478a1
+Found in base branch: develop
-Race condition in the ath_tx_aggr_sleep function in drivers/net/wireless/ath/ath9k/xmit.c in the Linux kernel before 3.13.7 allows remote attackers to cause a denial of service (system crash) via a large amount of network traffic that triggers certain list deletions. +All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. Steps to reproduce (provided by reporter Liyuan Chen): var lo = require('lodash'); function build_blank (n) { var ret = ""1"" for (var i = 0; i < n; i++) { ret += "" "" } return ret + ""1""; } var s = build_blank(50000) var time0 = Date.now(); lo.trim(s) var time_cost0 = Date.now() - time0; console.log(""time_cost0: "" + time_cost0) var time1 = Date.now(); lo.toNumber(s) var time_cost1 = Date.now() - time1; console.log(""time_cost1: "" + time_cost1) var time2 = Date.now(); lo.trimEnd(s) var time_cost2 = Date.now() - time2; console.log(""time_cost2: "" + time_cost2) -
Publish Date: 2014-04-01 -
URL: CVE-2014-2672
+Publish Date: 2021-02-15 +
URL: CVE-2020-28500
Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - - Attack Complexity: High + - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - - Availability Impact: High + - Availability Impact: Low
For more information on CVSS3 Scores, click here.- -
Type: Upgrade version
-Origin: https://nvd.nist.gov/vuln/detail/CVE-2014-2672
-Release Date: 2014-04-01
-Fix Resolution: 3.13.7
- - --
Julia Cartwright's fork of linux-stable-rt.git
-Library home page: https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git
-Found in HEAD commit: 5e23b7f9d2dd0154edd54986754eecd5b5308571
+",True,"CVE-2020-28500 (Medium) detected in lodash-1.0.2.tgz - ## CVE-2020-28500 - Medium Severity Vulnerability +Found in base branch: master
-A utility library delivering consistency, customization, performance, and extras.
+Library home page: https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz
+Path to dependency file: IoT-Smart-Gateway/web/package.json
+Path to vulnerable library: IoT-Smart-Gateway/web/node_modules/globule/node_modules/lodash/package.json
- /drivers/net/wireless/ath/ath9k/xmit.c
-
Found in HEAD commit: 085c2936b8c1a6952ca6f0db9eaf75ff33e478a1
+Found in base branch: develop
-Race condition in the ath_tx_aggr_sleep function in drivers/net/wireless/ath/ath9k/xmit.c in the Linux kernel before 3.13.7 allows remote attackers to cause a denial of service (system crash) via a large amount of network traffic that triggers certain list deletions. +All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. Steps to reproduce (provided by reporter Liyuan Chen): var lo = require('lodash'); function build_blank (n) { var ret = ""1"" for (var i = 0; i < n; i++) { ret += "" "" } return ret + ""1""; } var s = build_blank(50000) var time0 = Date.now(); lo.trim(s) var time_cost0 = Date.now() - time0; console.log(""time_cost0: "" + time_cost0) var time1 = Date.now(); lo.toNumber(s) var time_cost1 = Date.now() - time1; console.log(""time_cost1: "" + time_cost1) var time2 = Date.now(); lo.trimEnd(s) var time_cost2 = Date.now() - time2; console.log(""time_cost2: "" + time_cost2) -
Publish Date: 2014-04-01 -
URL: CVE-2014-2672
+Publish Date: 2021-02-15 +
URL: CVE-2020-28500
Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - - Attack Complexity: High + - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - - Availability Impact: High + - Availability Impact: Low
For more information on CVSS3 Scores, click here.-
Type: Upgrade version
-Origin: https://nvd.nist.gov/vuln/detail/CVE-2014-2672
-Release Date: 2014-04-01
-Fix Resolution: 3.13.7
+",0,cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file iot smart gateway web package json path to vulnerable library iot smart gateway web node modules globule node modules lodash package json dependency hierarchy gulp tgz root library vinyl fs tgz glob watcher tgz gaze tgz globule tgz x lodash tgz vulnerable library found in head commit a href found in base branch develop vulnerability details all versions of package lodash all versions of package org fujion webjars lodash are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions steps to reproduce provided by reporter liyuan chen var lo require lodash function build blank n var ret for var i i n i ret return ret var s build blank var date now lo trim s var time date now console log time time var date now lo tonumber s var time date now console log time time var date now lo trimend s var time date now console log time time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree gulp vinyl fs glob watcher gaze globule lodash isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails all versions of package lodash all versions of package org fujion webjars lodash are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions steps to reproduce provided by reporter liyuan chen var lo require function build blank n var ret for var i i n i ret return ret var s build blank var date now lo trim s var time date now console log time time var date now lo tonumber s var time date now console log time time var date now lo trimend s var time date now console log time time vulnerabilityurl ,0 +83599,7876177083.0,IssuesEvent,2018-06-25 23:29:17,Microsoft/vscode,https://api.github.com/repos/Microsoft/vscode,opened,Test: Webview persistence api,testplan-item,"Test for #49022 #48542 + +Complexity: 3 + +- [ ] Any OS +- [ ] Any OS + +**Summary** +A new api for webview serialization has been added. This allows webviews to both save their state between becoming hidden and visible, as well as saving their state when VS Code restarts + +**Test** +Try adding the two types of persistence to the new webview (you can use the [cat coding example](https://github.com/Microsoft/vscode-extension-samples/blob/master/webview-sample/) for this) + +Visibility persistence is implemented using a `getState` and `setState` method inside the webview itself: + +```ts +const vscode = acquireVsCodeApi(); +const state = vscode.getState(); +vscode.setState({ abc: 123 }) +``` + +Serialization uses [`WebviewPanelSerializer`](https://github.com/Microsoft/vscode/blob/f5300ced7fbd0193babf05ee37860bb2d160beb7/src/vs/vscode.d.ts#L5423) and [`registerWebviewPanelSerializer`](https://github.com/Microsoft/vscode/blob/f5300ced7fbd0193babf05ee37860bb2d160beb7/src/vs/vscode.d.ts#L5942) to register a handler in the extension that can revive a webview from its state. + + +**Things to test:** + +* Can you implement the two types of persistence? +* Does the API documentation make sense? (I'll be writing docs for it too but have not finished these yet) +* Are there cases where the new APIs may not work?",1.0,"Test: Webview persistence api - Test for #49022 #48542 + +Complexity: 3 + +- [ ] Any OS +- [ ] Any OS + +**Summary** +A new api for webview serialization has been added. This allows webviews to both save their state between becoming hidden and visible, as well as saving their state when VS Code restarts + +**Test** +Try adding the two types of persistence to the new webview (you can use the [cat coding example](https://github.com/Microsoft/vscode-extension-samples/blob/master/webview-sample/) for this) + +Visibility persistence is implemented using a `getState` and `setState` method inside the webview itself: + +```ts +const vscode = acquireVsCodeApi(); +const state = vscode.getState(); +vscode.setState({ abc: 123 }) +``` + +Serialization uses [`WebviewPanelSerializer`](https://github.com/Microsoft/vscode/blob/f5300ced7fbd0193babf05ee37860bb2d160beb7/src/vs/vscode.d.ts#L5423) and [`registerWebviewPanelSerializer`](https://github.com/Microsoft/vscode/blob/f5300ced7fbd0193babf05ee37860bb2d160beb7/src/vs/vscode.d.ts#L5942) to register a handler in the extension that can revive a webview from its state. + + +**Things to test:** + +* Can you implement the two types of persistence? +* Does the API documentation make sense? (I'll be writing docs for it too but have not finished these yet) +* Are there cases where the new APIs may not work?",0,test webview persistence api test for complexity any os any os summary a new api for webview serialization has been added this allows webviews to both save their state between becoming hidden and visible as well as saving their state when vs code restarts test try adding the two types of persistence to the new webview you can use the for this visibility persistence is implemented using a getstate and setstate method inside the webview itself ts const vscode acquirevscodeapi const state vscode getstate vscode setstate abc serialization uses and to register a handler in the extension that can revive a webview from its state things to test can you implement the two types of persistence does the api documentation make sense i ll be writing docs for it too but have not finished these yet are there cases where the new apis may not work ,0 +3272,6222643867.0,IssuesEvent,2017-07-10 09:40:10,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,opened,The field of `select 'a' from t1;` is wrong,compatibility,"Please answer these questions before submitting your issue. Thanks! + +1. What did you do? +``` +create table t1(a int); +insert into t value(1); +select 'a' from t1; +``` + +2. What did you expect to see? +``` ++---+ +| a | ++---+ +| a | ++---+ +``` + + +3. What did you see instead? +``` ++------+ +| 'a' | ++------+ +| a | ++------+ +``` + + +4. What version of TiDB are you using (`tidb-server -V`)? + +",True,"The field of `select 'a' from t1;` is wrong - Please answer these questions before submitting your issue. Thanks! + +1. What did you do? +``` +create table t1(a int); +insert into t value(1); +select 'a' from t1; +``` + +2. What did you expect to see? +``` ++---+ +| a | ++---+ +| a | ++---+ +``` + + +3. What did you see instead? +``` ++------+ +| 'a' | ++------+ +| a | ++------+ +``` + + +4. What version of TiDB are you using (`tidb-server -V`)? + +",1,the field of select a from is wrong please answer these questions before submitting your issue thanks what did you do create table a int insert into t value select a from what did you expect to see a a what did you see instead a a what version of tidb are you using tidb server v ,1 +11031,13065168365.0,IssuesEvent,2020-07-30 19:20:31,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,ifnull should not do type conversion?,type/compatibility,"This is a compatibility issue. + +1. What did you do? +If possible, provide a recipe for reproducing the error. + +```sql +use test; +create table t1 (id int unsigned); +insert into t1 values (NULL); +select ifnull(id, -1) from t1; +``` + +2. What did you expect to see? + +This is the output when executed on MySQL + +```sql +MariaDB [test]> create table t1 (id int unsigned); +Query OK, 0 rows affected (0.13 sec) + +MariaDB [test]> insert into t1 values (NULL); +Query OK, 1 row affected (0.00 sec) + +MariaDB [test]> select ifnull(id, -1) from t1; ++----------------+ +| ifnull(id, -1) | ++----------------+ +| -1 | ++----------------+ +1 row in set (0.00 sec) + +MariaDB [test]> + +``` + + +3. What did you see instead? + +```sql +TiDB-Server [test]> create table t1 (id int unsigned); +Query OK, 0 rows affected (0.01 sec) + +TiDB-Server [test]> insert into t1 values (NULL); +Query OK, 1 row affected (0.00 sec) + +TiDB-Server [test]> select ifnull(id, -1) from t1; ++----------------------+ +| ifnull(id, -1) | ++----------------------+ +| 18446744073709551615 | ++----------------------+ +1 row in set (0.00 sec) +``` + +since ifnull can also so use string as second parameter, convert from int to unsinged int seems not necessary. + +``` sql + +MySQL [test]> select ifnull(id, 'a') from t1; ++-----------------+ +| ifnull(id, 'a') | ++-----------------+ +| a | ++-----------------+ +1 row in set (0.01 sec) +``` + +4. What version of TiDB are you using (`tidb-server -V`)? + +Git Commit Hash: b1a3244e73e038589ab993ab02267528254ffcb2 + +",True,"ifnull should not do type conversion? - This is a compatibility issue. + +1. What did you do? +If possible, provide a recipe for reproducing the error. + +```sql +use test; +create table t1 (id int unsigned); +insert into t1 values (NULL); +select ifnull(id, -1) from t1; +``` + +2. What did you expect to see? + +This is the output when executed on MySQL + +```sql +MariaDB [test]> create table t1 (id int unsigned); +Query OK, 0 rows affected (0.13 sec) + +MariaDB [test]> insert into t1 values (NULL); +Query OK, 1 row affected (0.00 sec) + +MariaDB [test]> select ifnull(id, -1) from t1; ++----------------+ +| ifnull(id, -1) | ++----------------+ +| -1 | ++----------------+ +1 row in set (0.00 sec) + +MariaDB [test]> + +``` + + +3. What did you see instead? + +```sql +TiDB-Server [test]> create table t1 (id int unsigned); +Query OK, 0 rows affected (0.01 sec) + +TiDB-Server [test]> insert into t1 values (NULL); +Query OK, 1 row affected (0.00 sec) + +TiDB-Server [test]> select ifnull(id, -1) from t1; ++----------------------+ +| ifnull(id, -1) | ++----------------------+ +| 18446744073709551615 | ++----------------------+ +1 row in set (0.00 sec) +``` + +since ifnull can also so use string as second parameter, convert from int to unsinged int seems not necessary. + +``` sql + +MySQL [test]> select ifnull(id, 'a') from t1; ++-----------------+ +| ifnull(id, 'a') | ++-----------------+ +| a | ++-----------------+ +1 row in set (0.01 sec) +``` + +4. What version of TiDB are you using (`tidb-server -V`)? + +Git Commit Hash: b1a3244e73e038589ab993ab02267528254ffcb2 + +",1,ifnull should not do type conversion this is a compatibility issue what did you do if possible provide a recipe for reproducing the error sql use test create table id int unsigned insert into values null select ifnull id from what did you expect to see this is the output when executed on mysql sql mariadb create table id int unsigned query ok rows affected sec mariadb insert into values null query ok row affected sec mariadb select ifnull id from ifnull id row in set sec mariadb what did you see instead sql tidb server create table id int unsigned query ok rows affected sec tidb server insert into values null query ok row affected sec tidb server select ifnull id from ifnull id row in set sec since ifnull can also so use string as second parameter convert from int to unsinged int seems not necessary sql mysql select ifnull id a from ifnull id a a row in set sec what version of tidb are you using tidb server v git commit hash ,1 +10671,12609985513.0,IssuesEvent,2020-06-12 03:25:44,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,reopened,Feature request: Support CHECKSUM TABLE command,status/help-wanted type/compatibility type/enhancement,"When importing data into TiDB we want to be sure the data is correct. A convenient command in MySQL is CHECKSUM TABLE
-
+node-sass-4.11.0.tgz
Wrapper around libsass
+Library home page: https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz
+Path to dependency file: bootstrap/package.json
+Path to vulnerable library: bootstrap/node_modules/node-sass/package.json
+-### 2. What did you expect to see? (Required) +Dependency Hierarchy: + - :x: **node-sass-4.11.0.tgz** (Vulnerable Library) + -With MySQL 8.1.0: -``` -sql> CREATE TABLE t1 (id INT PRIMARY KEY, c1 TEXT DEFAULT 'foo'); -ERROR: 1101 (42000): BLOB, TEXT, GEOMETRY or JSON column 'c1' can't have a default value - -sql> CREATE TABLE t1 (id INT PRIMARY KEY, c1 TEXT DEFAULT ('foo')); -Query OK, 0 rows affected (0.0586 sec) - -sql> SHOW CREATE TABLE t1\G -*************************** 1. row *************************** - Table: t1 -Create Table: CREATE TABLE `t1` ( - `id` int NOT NULL, - `c1` text COLLATE utf8mb4_general_ci DEFAULT (_utf8mb4'foo'), - PRIMARY KEY (`id`) -) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_general_ci -1 row in set (0.0184 sec) - -sql> INSERT INTO t1(id) VALUES(1); -Query OK, 1 row affected (0.0076 sec) - -sql> TABLE t1; -+----+-----+ -| id | c1 | -+----+-----+ -| 1 | foo | -+----+-----+ -1 row in set (0.0009 sec) -``` -### 3. What did you see instead (Required) +
Found in base branch: v4-dev
+ + + ++ +In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file. -``` -sql> CREATE TABLE t1 (id INT PRIMARY KEY, c1 TEXT DEFAULT 'foo'); -ERROR: 1101 (42000): BLOB/TEXT/JSON column 'c1' can't have a default value +
Publish Date: 2018-12-03 +
URL: CVE-2018-19797
+ +-sql> CREATE TABLE t1 (id INT PRIMARY KEY, c1 TEXT DEFAULT ('foo')); -ERROR: 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your TiDB version for the right syntax to use line 1 column 59 near ""'foo'))"" +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: Required + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-... -``` +
Type: Upgrade version
+Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797
+Release Date: 2018-12-03
+Fix Resolution: libsass-3.6.0
-### 4. What is your TiDB version? (Required) + +
-Please answer these questions before submitting your issue. Thanks!
+node-sass-4.11.0.tgz
Wrapper around libsass
+Library home page: https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz
+Path to dependency file: bootstrap/package.json
+Path to vulnerable library: bootstrap/node_modules/node-sass/package.json
+-mysql> insert into t1 values (1,default,default); -Query OK, 1 row affected (0.01 sec) +Dependency Hierarchy: + - :x: **node-sass-4.11.0.tgz** (Vulnerable Library) +
Found in base branch: v4-dev
+ + + ++ +In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file. -``` -2. What did you expect to see? +
Publish Date: 2018-12-03 +
URL: CVE-2018-19797
+ +-``` -ERROR 3105 (HY000): The value specified for generated column 'b' in table 't1' is not allowed. -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: Required + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-3. What did you see instead? +
Type: Upgrade version
+Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797
+Release Date: 2018-12-03
+Fix Resolution: libsass-3.6.0
-``` -Query OK, 1 row affected (0.01 sec) -``` + +Tool for transforming CSS with JS plugins
+Library home page: https://registry.npmjs.org/postcss/-/postcss-4.1.16.tgz
+Path to dependency file: Michael-Hicks/package.json
+Path to vulnerable library: Michael-Hicks/node_modules/postcss/package.json
+-mysql> insert into t1 values (1,default,default) - -> on duplicate key update a=2, b=default; +Dependency Hierarchy: + - grunt-autoprefixer-3.0.4.tgz (Root Library) + - :x: **postcss-4.1.16.tgz** (Vulnerable Library) +
Found in HEAD commit: 057013f0e6c6f3c2e3d4b4a2ae5197624ad577d3
+Found in base branch: master
+ ++ +The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*). -``` -2. What did you expect to see? +
Publish Date: 2021-04-26 +
URL: CVE-2021-23382
+ +-``` -ERROR 3105 (HY000): The value specified for generated column 'b' in table 't1' is not allowed. -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: Low +
+For more information on CVSS3 Scores, click here. + +-3. What did you see instead? +
Type: Upgrade version
+Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382
+Release Date: 2021-04-26
+Fix Resolution: postcss - 8.2.13
-``` -Query OK, 1 row affected (0.01 sec) -``` + +Tool for transforming CSS with JS plugins
+Library home page: https://registry.npmjs.org/postcss/-/postcss-4.1.16.tgz
+Path to dependency file: Michael-Hicks/package.json
+Path to vulnerable library: Michael-Hicks/node_modules/postcss/package.json
+-Please answer these questions before submitting your issue. Thanks! +Dependency Hierarchy: + - grunt-autoprefixer-3.0.4.tgz (Root Library) + - :x: **postcss-4.1.16.tgz** (Vulnerable Library) +
Found in HEAD commit: 057013f0e6c6f3c2e3d4b4a2ae5197624ad577d3
+Found in base branch: master
+ ++ +The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*). -1. What did you do? -If possible, provide a recipe for reproducing the error. -``` -# ----- step 1: create table -CREATE TABLE `projects_send` ( - `entry` bigint(20) unsigned NOT NULL AUTO_INCREMENT COMMENT 'id', - `projectid` int(8) NOT NULL COMMENT 'ID', - `status` tinyint(1) unsigned NOT NULL DEFAULT '0' COMMENT '状态(0:not send;1:sent;2:back)', - `sendtime` datetime NOT NULL COMMENT 'senttime', - PRIMARY KEY (`entry`) USING BTREE, - KEY `idx_sendtime_status_projectid` (`sendtime`,`status`,`projectid`) USING BTREE, - KEY `idx_sendtime_status` (`sendtime`,`status`) USING BTREE -) ENGINE=InnoDB AUTO_INCREMENT=1 DEFAULT CHARSET=utf8; - -# ----- step 2: EXPLAIN on tidb & mysql - -#1. tidb & mysql same -mysql> -EXPLAIN select * from projects_send force index(idx_sendtime) where sendtime > ""2020-01-01""; -Key 'idx_sendtime' doesnt exist in table 'projects_send' - -#2. tidb & mysql same -mysql> -EXPLAIN select * from projects_send force index(idx_sendtime_status) where sendtime > ""2020-01-01""; -+----+-------------+---------------+-------+---------------------+---------------------+---------+------+------+--------------------------+ -| id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra | -+----+-------------+---------------+-------+---------------------+---------------------+---------+------+------+--------------------------+ -| 1 | SIMPLE | projects_send | index | idx_sendtime_status | idx_sendtime_status | 10 | NULL | 1 | Using where; Using index | -+----+-------------+---------------+-------+---------------------+---------------------+---------+------+------+--------------------------+ - -#3. tidb & mysql not same: tidb show index not exist -mysql> -EXPLAIN select * from projects_send force index(idx_sendtime_status_pro) where sendtime > ""2020-01-01""; -+----+-------------+---------------+-------+-------------------------------+-------------------------------+---------+------+------+--------------------------+ -| id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra | -+----+-------------+---------------+-------+-------------------------------+-------------------------------+---------+------+------+--------------------------+ -| 1 | SIMPLE | projects_send | index | idx_sendtime_status_projectid | idx_sendtime_status_projectid | 10 | NULL | 1 | Using where; Using index | -+----+-------------+---------------+-------+-------------------------------+-------------------------------+---------+------+------+--------------------------+ -1 row in set (0.14 sec) - -#on mysql the only one same prefix index idx_sendtime_status_pro can be used, but on tidb only the same name index can be use -``` -2. What did you expect to see? +
Publish Date: 2021-04-26 +
URL: CVE-2021-23382
+ ++Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Network + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: None + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: None + - Integrity Impact: None + - Availability Impact: Low +
+For more information on CVSS3 Scores, click here. + ++
Type: Upgrade version
+Origin: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382
+Release Date: 2021-04-26
+Fix Resolution: postcss - 8.2.13
-3. What did you see instead? + +IPython: Productive Interactive Computing
+Library home page: https://files.pythonhosted.org/packages/7c/6a/1f1365f4bf9fcb349fcaa5b61edfcefa721aa13ff37c5631296b12fab8e5/ipython-7.34.0-py3-none-any.whl
+Path to dependency file: /src/project/data-science/requirements.txt
+Path to vulnerable library: /src/project/data-science/requirements.txt
+-4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? -Release Version: v3.0.8 -Git Commit Hash: 8f13cf1449bd8903ff465a4f12ed89ecbac858a4 -Git Branch: HEAD -UTC Build Time: 2019-12-31 11:14:59 -GoVersion: go version go1.13 linux/amd64 -Race Enabled: false -TiKV Min Version: v3.0.0-60965b006877ca7234adaced7890d7b029ed1306 -Check Table Before Drop: false -",True,"force index behaviour not same as mysql - ## Bug Report +Dependency Hierarchy: + - jupyter-1.0.0-py2.py3-none-any.whl (Root Library) + - jupyter_console-6.6.3-py3-none-any.whl + - :x: **ipython-7.34.0-py3-none-any.whl** (Vulnerable Library) +
Found in HEAD commit: 92e57f9e81da15812523bf929f8ad33bdae5e967
+Found in base branch: main
+ ++ +IPython (Interactive Python) is a command shell for interactive computing in multiple programming languages, originally developed for the Python programming language. Versions prior to 8.1.0 are subject to a command injection vulnerability with very specific prerequisites. This vulnerability requires that the function `IPython.utils.terminal.set_term_title` be called on Windows in a Python environment where ctypes is not available. The dependency on `ctypes` in `IPython.utils._process_win32` prevents the vulnerable code from ever being reached in the ipython binary. However, as a library that could be used by another tool `set_term_title` could be called and hence introduce a vulnerability. Should an attacker get untrusted input to an instance of this function they would be able to inject shell commands as current process and limited to the scope of the current process. Users of ipython as a library are advised to upgrade. Users unable to upgrade should ensure that any calls to the `IPython.utils.terminal.set_term_title` function are done with trusted or filtered input. -Please answer these questions before submitting your issue. Thanks! +
Publish Date: 2023-02-10 +
URL: CVE-2023-24816
+ +-1. What did you do? -If possible, provide a recipe for reproducing the error. -``` -# ----- step 1: create table -CREATE TABLE `projects_send` ( - `entry` bigint(20) unsigned NOT NULL AUTO_INCREMENT COMMENT 'id', - `projectid` int(8) NOT NULL COMMENT 'ID', - `status` tinyint(1) unsigned NOT NULL DEFAULT '0' COMMENT '状态(0:not send;1:sent;2:back)', - `sendtime` datetime NOT NULL COMMENT 'senttime', - PRIMARY KEY (`entry`) USING BTREE, - KEY `idx_sendtime_status_projectid` (`sendtime`,`status`,`projectid`) USING BTREE, - KEY `idx_sendtime_status` (`sendtime`,`status`) USING BTREE -) ENGINE=InnoDB AUTO_INCREMENT=1 DEFAULT CHARSET=utf8; - -# ----- step 2: EXPLAIN on tidb & mysql - -#1. tidb & mysql same -mysql> -EXPLAIN select * from projects_send force index(idx_sendtime) where sendtime > ""2020-01-01""; -Key 'idx_sendtime' doesnt exist in table 'projects_send' - -#2. tidb & mysql same -mysql> -EXPLAIN select * from projects_send force index(idx_sendtime_status) where sendtime > ""2020-01-01""; -+----+-------------+---------------+-------+---------------------+---------------------+---------+------+------+--------------------------+ -| id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra | -+----+-------------+---------------+-------+---------------------+---------------------+---------+------+------+--------------------------+ -| 1 | SIMPLE | projects_send | index | idx_sendtime_status | idx_sendtime_status | 10 | NULL | 1 | Using where; Using index | -+----+-------------+---------------+-------+---------------------+---------------------+---------+------+------+--------------------------+ - -#3. tidb & mysql not same: tidb show index not exist -mysql> -EXPLAIN select * from projects_send force index(idx_sendtime_status_pro) where sendtime > ""2020-01-01""; -+----+-------------+---------------+-------+-------------------------------+-------------------------------+---------+------+------+--------------------------+ -| id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra | -+----+-------------+---------------+-------+-------------------------------+-------------------------------+---------+------+------+--------------------------+ -| 1 | SIMPLE | projects_send | index | idx_sendtime_status_projectid | idx_sendtime_status_projectid | 10 | NULL | 1 | Using where; Using index | -+----+-------------+---------------+-------+-------------------------------+-------------------------------+---------+------+------+--------------------------+ -1 row in set (0.14 sec) - -#on mysql the only one same prefix index idx_sendtime_status_pro can be used, but on tidb only the same name index can be use -``` -2. What did you expect to see? +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Local + - Attack Complexity: High + - Privileges Required: None + - User Interaction: Required + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + ++
Type: Upgrade version
+Origin: https://www.cve.org/CVERecord?id=CVE-2023-24816
+Release Date: 2023-02-10
+Fix Resolution: ipython - 8.10.0
+ +IPython: Productive Interactive Computing
+Library home page: https://files.pythonhosted.org/packages/7c/6a/1f1365f4bf9fcb349fcaa5b61edfcefa721aa13ff37c5631296b12fab8e5/ipython-7.34.0-py3-none-any.whl
+Path to dependency file: /src/project/data-science/requirements.txt
+Path to vulnerable library: /src/project/data-science/requirements.txt
++Dependency Hierarchy: + - jupyter-1.0.0-py2.py3-none-any.whl (Root Library) + - jupyter_console-6.6.3-py3-none-any.whl + - :x: **ipython-7.34.0-py3-none-any.whl** (Vulnerable Library) +
Found in HEAD commit: 92e57f9e81da15812523bf929f8ad33bdae5e967
+Found in base branch: main
+ ++ +IPython (Interactive Python) is a command shell for interactive computing in multiple programming languages, originally developed for the Python programming language. Versions prior to 8.1.0 are subject to a command injection vulnerability with very specific prerequisites. This vulnerability requires that the function `IPython.utils.terminal.set_term_title` be called on Windows in a Python environment where ctypes is not available. The dependency on `ctypes` in `IPython.utils._process_win32` prevents the vulnerable code from ever being reached in the ipython binary. However, as a library that could be used by another tool `set_term_title` could be called and hence introduce a vulnerability. Should an attacker get untrusted input to an instance of this function they would be able to inject shell commands as current process and limited to the scope of the current process. Users of ipython as a library are advised to upgrade. Users unable to upgrade should ensure that any calls to the `IPython.utils.terminal.set_term_title` function are done with trusted or filtered input. -4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? -Release Version: v3.0.8 -Git Commit Hash: 8f13cf1449bd8903ff465a4f12ed89ecbac858a4 -Git Branch: HEAD -UTC Build Time: 2019-12-31 11:14:59 -GoVersion: go version go1.13 linux/amd64 -Race Enabled: false -TiKV Min Version: v3.0.0-60965b006877ca7234adaced7890d7b029ed1306 -Check Table Before Drop: false -",1,force index behaviour not same as mysql bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error step create table create table projects send entry bigint unsigned not null auto increment comment id projectid int not null comment id status tinyint unsigned not null default comment 状态 not send sent back sendtime datetime not null comment senttime primary key entry using btree key idx sendtime status projectid sendtime status projectid using btree key idx sendtime status sendtime status using btree engine innodb auto increment default charset step explain on tidb mysql tidb mysql same mysql explain select from projects send force index idx sendtime where sendtime key idx sendtime doesnt exist in table projects send tidb mysql same mysql explain select from projects send force index idx sendtime status where sendtime id select type table type possible keys key key len ref rows extra simple projects send index idx sendtime status idx sendtime status null using where using index tidb mysql not same tidb show index not exist mysql explain select from projects send force index idx sendtime status pro where sendtime id select type table type possible keys key key len ref rows extra simple projects send index idx sendtime status projectid idx sendtime status projectid null using where using index row in set sec on mysql the only one same prefix index idx sendtime status pro can be used but on tidb only the same name index can be use what did you expect to see what did you see instead what version of tidb are you using tidb server v or run select tidb version on tidb release version git commit hash git branch head utc build time goversion go version linux race enabled false tikv min version check table before drop false ,1 -11624,13671944268.0,IssuesEvent,2020-09-29 07:48:08,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,_utf8 not supported,component/parser type/compatibility,"Please answer these questions before submitting your issue. Thanks! +
Publish Date: 2023-02-10 +
URL: CVE-2023-24816
+ +-1. What did you do? -If possible, provide a recipe for reproducing the error. +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Local + - Attack Complexity: High + - Privileges Required: None + - User Interaction: Required + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-```sql -select left(_utf8 0xD0B0D0B1D0B2,1) -``` +
Type: Upgrade version
+Origin: https://www.cve.org/CVERecord?id=CVE-2023-24816
+Release Date: 2023-02-10
+Fix Resolution: ipython - 8.10.0
-2. What did you expect to see? + +
-3. What did you see instead?
+async-2.0.0.tgz
Higher-order functions and common patterns for asynchronous code
+Library home page: https://registry.npmjs.org/async/-/async-2.0.0.tgz
+Path to dependency file: /package.json
+Path to vulnerable library: /node_modules/socketcluster-server/node_modules/async/package.json,/node_modules/sc-broker-cluster/node_modules/async/package.json,/node_modules/socketcluster/node_modules/async/package.json
+-4. What version of TiDB are you using (`tidb-server -V`)? +Dependency Hierarchy: + - remotedev-rn-debugger-0.8.3.tgz (Root Library) + - remotedev-server-0.2.7.tgz + - socketcluster-8.0.2.tgz + - :x: **async-2.0.0.tgz** (Vulnerable Library) +
Higher-order functions and common patterns for asynchronous code
+Library home page: https://registry.npmjs.org/async/-/async-2.6.2.tgz
+Path to dependency file: /package.json
+Path to vulnerable library: /node_modules/async/package.json
+-1. What did you do? -If possible, provide a recipe for reproducing the error. +Dependency Hierarchy: + - react-native-0.55.4.tgz (Root Library) + - metro-0.30.2.tgz + - :x: **async-2.6.2.tgz** (Vulnerable Library) +
Found in base branch: stable
+ + + ++ +In Async before 2.6.4 and 3.x before 3.2.2, a malicious user can obtain privileges via the mapValues() method, aka lib/internal/iterator.js createObjectIterator prototype pollution. -2. What did you expect to see? +
Publish Date: 2022-04-06 +
URL: CVE-2021-43138
+ +-``` -+-----------------------------------------------+ -| (1.175494351E-37 div 1.7976931348623157E+308) | -+-----------------------------------------------+ -| 0 | -+-----------------------------------------------+ -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Local + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: Required + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-3. What did you see instead? +
Type: Upgrade version
+Origin: https://nvd.nist.gov/vuln/detail/CVE-2021-43138
+Release Date: 2022-04-06
+Fix Resolution (async): 2.6.4
+Direct dependency fix Resolution (react-native): 0.65.0
-``` -ERROR 1690 (22003): %s value is out of range in '%s' -``` + +
-",1, div scientific notation values out of range please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error sql select div what did you expect to see div what did you see instead error s value is out of range in s what version of tidb are you using tidb server v ,1
-451606,13039229324.0,IssuesEvent,2020-07-28 16:24:07,googleapis/release-please,https://api.github.com/repos/googleapis/release-please,closed,Cannot read property 'sha' of undefined (Java),priority: p2 type: bug,"We're seeing a bunch of these in the `repo-automation-bots` logs:
+async-2.0.0.tgz
Higher-order functions and common patterns for asynchronous code
+Library home page: https://registry.npmjs.org/async/-/async-2.0.0.tgz
+Path to dependency file: /package.json
+Path to vulnerable library: /node_modules/socketcluster-server/node_modules/async/package.json,/node_modules/sc-broker-cluster/node_modules/async/package.json,/node_modules/socketcluster/node_modules/async/package.json
+-It looks like it's possible for the number of commits in this situation to be 0: +Dependency Hierarchy: + - remotedev-rn-debugger-0.8.3.tgz (Root Library) + - remotedev-server-0.2.7.tgz + - socketcluster-8.0.2.tgz + - :x: **async-2.0.0.tgz** (Vulnerable Library) +
Higher-order functions and common patterns for asynchronous code
+Library home page: https://registry.npmjs.org/async/-/async-2.6.2.tgz
+Path to dependency file: /package.json
+Path to vulnerable library: /node_modules/async/package.json
+-In this case, the size of the value returned from `this.commits` is 0. ",1.0,"Cannot read property 'sha' of undefined (Java) - We're seeing a bunch of these in the `repo-automation-bots` logs: +Dependency Hierarchy: + - react-native-0.55.4.tgz (Root Library) + - metro-0.30.2.tgz + - :x: **async-2.6.2.tgz** (Vulnerable Library) +
Found in base branch: stable
+ + + ++ +In Async before 2.6.4 and 3.x before 3.2.2, a malicious user can obtain privileges via the mapValues() method, aka lib/internal/iterator.js createObjectIterator prototype pollution. -It looks like it's possible for the number of commits in this situation to be 0: +
Publish Date: 2022-04-06 +
URL: CVE-2021-43138
+ +-```js -if (this.snapshot) { - const latestCommit = ( - await this.commits(latestTag ? latestTag.sha : undefined, 1, true) - )[0]; - prSHA = latestCommit.sha; - } -``` +Base Score Metrics: +- Exploitability Metrics: + - Attack Vector: Local + - Attack Complexity: Low + - Privileges Required: None + - User Interaction: Required + - Scope: Unchanged +- Impact Metrics: + - Confidentiality Impact: High + - Integrity Impact: High + - Availability Impact: High +
+For more information on CVSS3 Scores, click here. + +-In this case, the size of the value returned from `this.commits` is 0. ",0,cannot read property sha of undefined java we re seeing a bunch of these in the repo automation bots logs typeerror cannot read property sha of undefined at javayoshi run workspace node modules release please build src releasers java yoshi js it looks like it s possible for the number of commits in this situation to be js if this snapshot const latestcommit await this commits latesttag latesttag sha undefined true prsha latestcommit sha in this case the size of the value returned from this commits is ,0 -11111,13131607739.0,IssuesEvent,2020-08-06 17:19:47,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,error message is not compatible with MySQL,good-first-issue status/help-wanted type/compatibility,"## Bug Report - -Please answer these questions before submitting your issue. Thanks! +
Type: Upgrade version
+Origin: https://nvd.nist.gov/vuln/detail/CVE-2021-43138
+Release Date: 2022-04-06
+Fix Resolution (async): 2.6.4
+Direct dependency fix Resolution (react-native): 0.65.0
-1. What did you do? -``` sql -create table t(a int); -select a as b from t group by values(b); -``` + +Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/4.4.0/system.security.cryptography.xml.4.4.0.nupkg
-+### What happened -
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section ""Details"" below to see if there is a version of transitive dependency where vulnerability is fixed.
+### What you expected to happen -## Details +The workflow status field of the task definition page does not support Chinese. -+_No response_ +### Version -### microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+dev -Core components of ASP.NET Core Kestrel cross-platform web server.
-Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
-Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
-Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
-+### Are you willing to submit PR? -Dependency Hierarchy: - - dotnetcore-sample-1.0.0 (Root Library) - - microsoft.aspnetcore.2.0.0.nupkg - - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg - - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) +- [X] Yes I am willing to submit a PR! +### Code of Conduct -### microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
+- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct) +",1.0,"[Bug][UI Next][V1.0.0-Alpha] The workflow status field of the task definition page does not support Chinese. - ### Search before asking -ASP.NET Core basic middleware for supporting HTTP method overrides. Includes: -* X-Forwarded-* header...
-Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
-Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
-Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.httpoverrides/2.0.0/microsoft.aspnetcore.httpoverrides.2.0.0.nupkg
-+- [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues. -Dependency Hierarchy: - - dotnetcore-sample-1.0.0 (Root Library) - - microsoft.aspnetcore.2.0.0.nupkg - - microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg - - :x: **microsoft.aspnetcore.httpoverrides.2.0.0.nupkg** (Vulnerable Library) +### What happened -
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
-Found in base branch: main
- + - -### Vulnerability Details -- -ASP.NET Core 1.0. 1.1, and 2.0 allow an elevation of privilege vulnerability due to how web applications that are created from templates validate web requests, aka ""ASP.NET Core Elevation Of Privilege Vulnerability"". +### What you expected to happen -
Publish Date: 2018-03-14 -
URL: CVE-2018-0787
- +The workflow status field of the task definition page does not support Chinese. - +### How to reproduce -### CVSS 3 Score Details (8.8) -+Support chinese. -Base Score Metrics: -- Exploitability Metrics: - - Attack Vector: Network - - Attack Complexity: Low - - Privileges Required: None - - User Interaction: Required - - Scope: Unchanged -- Impact Metrics: - - Confidentiality Impact: High - - Integrity Impact: High - - Availability Impact: High -
-For more information on CVSS3 Scores, click here. - +### Anything else - +_No response_ -### Suggested Fix -+### Version -
Type: Upgrade version
-Release Date: 2018-03-14
-Fix Resolution: Microsoft.AspNetCore.HttpOverrides - 2.0.2, Microsoft.AspNetCore.Server.Kestrel.Core - 2.0.2
+dev - +### Are you willing to submit PR? - +- [X] Yes I am willing to submit a PR! -+### Is your feature request related to a problem? +_No response_ -### microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+### Describe the feature you'd like -Core components of ASP.NET Core Kestrel cross-platform web server.
-Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
-Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
-Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
-+fill(expr) + +**No filling**: NONE (default filling mode). +**Value filling**: Filling with a fixed value, where you need to specify the value to fill. For example: FILL(VALUE, 1.23). It is important to note that the final filling value depends on the type of the corresponding column. For example, if the column is of type INT, the filling value will be 1. +**Previous filling**: Filling with the previous non-NULL value. For example: FILL(PREV). +**Null filling**: Filling with NULL values. For example: FILL(NULL). +**Linear filling**: Filling with linear interpolation based on the nearest non-NULL values before and after. For example: FILL(LINEAR). +**Next filling**: Filling with the next non-NULL value. For example: FILL(NEXT). -Dependency Hierarchy: - - dotnetcore-sample-1.0.0 (Root Library) - - microsoft.aspnetcore.2.0.0.nupkg - - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg - - :x: **microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg** (Vulnerable Library) +### Describe implementation you've considered +_No response_ -### microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
+### Documentation, Adoption, Use Case, Migration Strategy -Transport abstractions for the ASP.NET Core Kestrel cross-platform web server.
-Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
-Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
-Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.transport.abstractions/2.0.0/microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg
-+```Markdown +Example: + +CREATE TABLE temperature ( + timestamp TIMESTAMP, + value FLOAT +); + +INSERT INTO temperature VALUES ('2023-08-29 00:00:00', 25.0); +INSERT INTO temperature VALUES ('2023-08-29 01:00:00', 27.0); +INSERT INTO temperature VALUES ('2023-08-29 02:00:00', NULL); +INSERT INTO temperature VALUES ('2023-08-29 03:00:00', NULL); +INSERT INTO temperature VALUES ('2023-08-29 04:00:00', 28.0); +INSERT INTO temperature VALUES ('2023-08-29 05:00:00', 30.0); + +SELECT timestamp, FILL(LINEAR) AS value FROM temperature; +``` -Dependency Hierarchy: - - dotnetcore-sample-1.0.0 (Root Library) - - microsoft.aspnetcore.2.0.0.nupkg - - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg - - microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg - - :x: **microsoft.aspnetcore.server.kestrel.transport.abstractions.2.0.0.nupkg** (Vulnerable Library) +### Additional information -### microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg
+_No response_",1.0,"[Feature Request]: filling query - ### Is there an existing issue for the same feature request? -Libuv transport for the ASP.NET Core Kestrel cross-platform web server.
-Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg
-Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
-Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.transport.libuv/2.0.0/microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg
-+- [X] I have checked the existing issues. -Dependency Hierarchy: - - dotnetcore-sample-1.0.0 (Root Library) - - microsoft.aspnetcore.2.0.0.nupkg - - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg - - :x: **microsoft.aspnetcore.server.kestrel.transport.libuv.2.0.0.nupkg** (Vulnerable Library) +### Is your feature request related to a problem? +_No response_ -
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
-Found in base branch: main
- +### Describe the feature you'd like - +fill(expr) + +**No filling**: NONE (default filling mode). +**Value filling**: Filling with a fixed value, where you need to specify the value to fill. For example: FILL(VALUE, 1.23). It is important to note that the final filling value depends on the type of the corresponding column. For example, if the column is of type INT, the filling value will be 1. +**Previous filling**: Filling with the previous non-NULL value. For example: FILL(PREV). +**Null filling**: Filling with NULL values. For example: FILL(NULL). +**Linear filling**: Filling with linear interpolation based on the nearest non-NULL values before and after. For example: FILL(LINEAR). +**Next filling**: Filling with the next non-NULL value. For example: FILL(NEXT). -### Vulnerability Details -- -A vulnerability was discovered in versions 2.x of ASP.NET Core where a specially crafted request can cause excess resource consumption in Kestrel. +### Describe implementation you've considered -
Publish Date: 2018-05-08 -
URL: WS-2018-0608
- +_No response_ - +### Documentation, Adoption, Use Case, Migration Strategy -### CVSS 3 Score Details (7.5) -+```Markdown +Example: + +CREATE TABLE temperature ( + timestamp TIMESTAMP, + value FLOAT +); + +INSERT INTO temperature VALUES ('2023-08-29 00:00:00', 25.0); +INSERT INTO temperature VALUES ('2023-08-29 01:00:00', 27.0); +INSERT INTO temperature VALUES ('2023-08-29 02:00:00', NULL); +INSERT INTO temperature VALUES ('2023-08-29 03:00:00', NULL); +INSERT INTO temperature VALUES ('2023-08-29 04:00:00', 28.0); +INSERT INTO temperature VALUES ('2023-08-29 05:00:00', 30.0); + +SELECT timestamp, FILL(LINEAR) AS value FROM temperature; +``` -Base Score Metrics: -- Exploitability Metrics: - - Attack Vector: Network - - Attack Complexity: Low - - Privileges Required: None - - User Interaction: None - - Scope: Unchanged -- Impact Metrics: - - Confidentiality Impact: None - - Integrity Impact: None - - Availability Impact: High -
-For more information on CVSS3 Scores, click here. - - +### Additional information -### Suggested Fix -
+_No response_",0, filling query is there an existing issue for the same feature request i have checked the existing issues is your feature request related to a problem no response describe the feature you d like fill expr no filling none default filling mode value filling filling with a fixed value where you need to specify the value to fill for example fill value it is important to note that the final filling value depends on the type of the corresponding column for example if the column is of type int the filling value will be previous filling filling with the previous non null value for example fill prev null filling filling with null values for example fill null linear filling filling with linear interpolation based on the nearest non null values before and after for example fill linear next filling filling with the next non null value for example fill next describe implementation you ve considered no response documentation adoption use case migration strategy markdown example create table temperature timestamp timestamp value float insert into temperature values insert into temperature values insert into temperature values null insert into temperature values null insert into temperature values insert into temperature values select timestamp fill linear as value from temperature additional information no response ,0
+7013,9301980842.0,IssuesEvent,2019-03-24 04:20:05,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,`select constants which have a plus prefix` is not same with MySQL,help wanted type/compatibility,"## Bug Report
+
+Please answer these questions before submitting your issue. Thanks!
+
+1. What did you do?
+If possible, provide a recipe for reproducing the error.
+In TiDB:
+mysql> select +1000;
++-------+
+| +1000 |
++-------+
+| 1000 |
++-------+
+1 row in set (0.00 sec)
+In MySQL:
+mysql> select +1000;
++------+
+| 1000 |
++------+
+| 1000 |
++------+
+1 row in set (0.00 sec)
+
+
+2. What did you expect to see?
+No plus display before column name.
+
+3. What did you see instead?
+A plus before column name.
+
+4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
+mysql> select tidb_version();
++---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+| tidb_version() |
++---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+| Release Version: v3.0.0-beta-202-g85b1cd7-dirty
+Git Commit Hash: 85b1cd7aa2d21119e39d4cc9a317557aaf7aa13c
+Git Branch: master
+UTC Build Time: 2019-03-12 12:17:58
+GoVersion: go version go1.12 linux/amd64
+Race Enabled: false
+TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e
+Check Table Before Drop: false |
++---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+1 row in set (0.00 sec)
+",True,"`select constants which have a plus prefix` is not same with MySQL - ## Bug Report
+
+Please answer these questions before submitting your issue. Thanks!
+
+1. What did you do?
+If possible, provide a recipe for reproducing the error.
+In TiDB:
+mysql> select +1000;
++-------+
+| +1000 |
++-------+
+| 1000 |
++-------+
+1 row in set (0.00 sec)
+In MySQL:
+mysql> select +1000;
++------+
+| 1000 |
++------+
+| 1000 |
++------+
+1 row in set (0.00 sec)
+
+
+2. What did you expect to see?
+No plus display before column name.
+
+3. What did you see instead?
+A plus before column name.
+
+4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
+mysql> select tidb_version();
++---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+| tidb_version() |
++---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+| Release Version: v3.0.0-beta-202-g85b1cd7-dirty
+Git Commit Hash: 85b1cd7aa2d21119e39d4cc9a317557aaf7aa13c
+Git Branch: master
+UTC Build Time: 2019-03-12 12:17:58
+GoVersion: go version go1.12 linux/amd64
+Race Enabled: false
+TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e
+Check Table Before Drop: false |
++---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+1 row in set (0.00 sec)
+",1, select constants which have a plus prefix is not same with mysql bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error in tidb mysql select row in set sec in mysql mysql select row in set sec what did you expect to see no plus display before column name what did you see instead a plus before column name what version of tidb are you using tidb server v or run select tidb version on tidb mysql select tidb version tidb version release version beta dirty git commit hash git branch master utc build time goversion go version linux race enabled false tikv min version alpha check table before drop false row in set sec ,1
+10287,8872808847.0,IssuesEvent,2019-01-11 16:20:24,terraform-providers/terraform-provider-aws,https://api.github.com/repos/terraform-providers/terraform-provider-aws,closed,Resource aws_emr_cluster doesn't support import,enhancement service/emr,"
+
+### Community Note
+
+* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
+* Please do not leave ""+1"" or ""me too"" comments, they generate extra noise for issue followers and do not help prioritize the request
+* If you are interested in working on this issue or have submitted a pull request, please leave a comment
+
+
+
+### Description
+We'd like to import our existing EMR clusters into terraform
+
+### New or Affected Resource(s)
+
+* aws_emr_cluster
+
+### Potential Terraform Configuration
+
+terraform import aws_emr_cluster.emr-test-cluster j-XXXXXXXXXXXXX
+
+### References
+
+
+
+* https://www.terraform.io/docs/providers/aws/r/emr_cluster.html
+",1.0,"Resource aws_emr_cluster doesn't support import -
+
+### Community Note
+
+* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
+* Please do not leave ""+1"" or ""me too"" comments, they generate extra noise for issue followers and do not help prioritize the request
+* If you are interested in working on this issue or have submitted a pull request, please leave a comment
+
+
+
+### Description
+We'd like to import our existing EMR clusters into terraform
+
+### New or Affected Resource(s)
+
+* aws_emr_cluster
+
+### Potential Terraform Configuration
+
+terraform import aws_emr_cluster.emr-test-cluster j-XXXXXXXXXXXXX
+
+### References
+
+
+
+* https://www.terraform.io/docs/providers/aws/r/emr_cluster.html
+",0,resource aws emr cluster doesn t support import community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description we d like to import our existing emr clusters into terraform new or affected resource s aws emr cluster potential terraform configuration terraform import aws emr cluster emr test cluster j xxxxxxxxxxxxx references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example ,0
+175215,13540303183.0,IssuesEvent,2020-09-16 14:30:21,microsoft/botframework-sdk,https://api.github.com/repos/microsoft/botframework-sdk,closed,"DCR: As an NLU engineer, I would like to be able to run batch tests scripts against my language models to validate the utterances and entities my bot needs to work get resolved as expected",NLU Testing,"## Issue
+The natural language understanding models used by a bot evolve independently of the underlying code, the LUIS portal provides a batch testing UI but it is not scriptable and can't be used in continuous integration or continuous delivery.
+
+## Proposed change
+We need a batch testing API that can be invoked and asserted from the Azure CLI to ensure the key utterances and entities are still resolved after the model changes.
+
+The API should take a json file that follows the [batch testing format](https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-concept-batch-test#batch-file-format) define by LUIS and output the results in a document format that can be parsed and asserted through automated tools.",1.0,"DCR: As an NLU engineer, I would like to be able to run batch tests scripts against my language models to validate the utterances and entities my bot needs to work get resolved as expected - ## Issue
+The natural language understanding models used by a bot evolve independently of the underlying code, the LUIS portal provides a batch testing UI but it is not scriptable and can't be used in continuous integration or continuous delivery.
+
+## Proposed change
+We need a batch testing API that can be invoked and asserted from the Azure CLI to ensure the key utterances and entities are still resolved after the model changes.
+
+The API should take a json file that follows the [batch testing format](https://docs.microsoft.com/en-us/azure/cognitive-services/luis/luis-concept-batch-test#batch-file-format) define by LUIS and output the results in a document format that can be parsed and asserted through automated tools.",0,dcr as an nlu engineer i would like to be able to run batch tests scripts against my language models to validate the utterances and entities my bot needs to work get resolved as expected issue the natural language understanding models used by a bot evolve independently of the underlying code the luis portal provides a batch testing ui but it is not scriptable and can t be used in continuous integration or continuous delivery proposed change we need a batch testing api that can be invoked and asserted from the azure cli to ensure the key utterances and entities are still resolved after the model changes the api should take a json file that follows the define by luis and output the results in a document format that can be parsed and asserted through automated tools ,0
+7145,9430605819.0,IssuesEvent,2019-04-12 09:25:49,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,json_key is not same with MySQL,type/compatibility,"## Bug Report
+
+Please answer these questions before submitting your issue. Thanks!
+
+1. What did you do?
+If possible, provide a recipe for reproducing the error.
+```sql
+select json_keys(NULL);
+select json_keys(NULL, '$.b');
+select json_keys(NULL, NULL);
+select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', '$.a');
+select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL);
+```
+
+2. What did you expect to see?
+## In MySQL
+```SQL
+mysql> select version();
++-------------------------+
+| version() |
++-------------------------+
+| 5.7.25-0ubuntu0.18.04.2 |
++-------------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL);
++-----------------+
+| json_keys(NULL) |
++-----------------+
+| NULL |
++-----------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL, '$.b');
++------------------------+
+| json_keys(NULL, '$.b') |
++------------------------+
+| NULL |
++------------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL, NULL);
++-----------------------+
+| json_keys(NULL, NULL) |
++-----------------------+
+| NULL |
++-----------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', '$.a');
++---------------------------------------------------------+
+| json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', '$.a') |
++---------------------------------------------------------+
+| NULL |
++---------------------------------------------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL);
++--------------------------------------------------------+
+| json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL) |
++--------------------------------------------------------+
+| NULL |
++--------------------------------------------------------+
+1 row in set (0.00 sec)
+
+mysql>
+```
+
+
+3. What did you see instead?
+## In TiDB
+```SQL
+
+mysql> select json_keys(NULL);
++-----------------+
+| json_keys(NULL) |
++-----------------+
+| NULL |
++-----------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL, '$.b');
++------------------------+
+| json_keys(NULL, '$.b') |
++------------------------+
+| NULL |
++------------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL, NULL);
++-----------------------+
+| json_keys(NULL, NULL) |
++-----------------------+
+| NULL |
++-----------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', '$.a');
+ERROR 3146 (22032): Invalid data type for JSON data
+mysql> select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL);
++--------------------------------------------------------+
+| json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL) |
++--------------------------------------------------------+
+| NULL |
++--------------------------------------------------------+
+1 row in set (0.01 sec)
+
+mysql> select tidb_version()\G
+*************************** 1. row ***************************
+tidb_version(): Release Version: v3.0.0-beta-321-gc6a757b24
+Git Commit Hash: c6a757b2435965a72ada9d61ea8b7a367bd89b5f
+Git Branch: master
+UTC Build Time: 2019-04-02 02:38:04
+GoVersion: go version go1.12.1 linux/amd64
+Race Enabled: false
+TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e
+Check Table Before Drop: false
+1 row in set (0.00 sec)
+
+mysql>
+```",True,"json_key is not same with MySQL - ## Bug Report
+
+Please answer these questions before submitting your issue. Thanks!
+
+1. What did you do?
+If possible, provide a recipe for reproducing the error.
+```sql
+select json_keys(NULL);
+select json_keys(NULL, '$.b');
+select json_keys(NULL, NULL);
+select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', '$.a');
+select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL);
+```
+
+2. What did you expect to see?
+## In MySQL
+```SQL
+mysql> select version();
++-------------------------+
+| version() |
++-------------------------+
+| 5.7.25-0ubuntu0.18.04.2 |
++-------------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL);
++-----------------+
+| json_keys(NULL) |
++-----------------+
+| NULL |
++-----------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL, '$.b');
++------------------------+
+| json_keys(NULL, '$.b') |
++------------------------+
+| NULL |
++------------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL, NULL);
++-----------------------+
+| json_keys(NULL, NULL) |
++-----------------------+
+| NULL |
++-----------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', '$.a');
++---------------------------------------------------------+
+| json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', '$.a') |
++---------------------------------------------------------+
+| NULL |
++---------------------------------------------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL);
++--------------------------------------------------------+
+| json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL) |
++--------------------------------------------------------+
+| NULL |
++--------------------------------------------------------+
+1 row in set (0.00 sec)
+
+mysql>
+```
+
+
+3. What did you see instead?
+## In TiDB
+```SQL
+
+mysql> select json_keys(NULL);
++-----------------+
+| json_keys(NULL) |
++-----------------+
+| NULL |
++-----------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL, '$.b');
++------------------------+
+| json_keys(NULL, '$.b') |
++------------------------+
+| NULL |
++------------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys(NULL, NULL);
++-----------------------+
+| json_keys(NULL, NULL) |
++-----------------------+
+| NULL |
++-----------------------+
+1 row in set (0.00 sec)
+
+mysql> select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', '$.a');
+ERROR 3146 (22032): Invalid data type for JSON data
+mysql> select json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL);
++--------------------------------------------------------+
+| json_keys('{""a"": 1, ""b"": {""e"": ""foo"", ""b"": 3}}', NULL) |
++--------------------------------------------------------+
+| NULL |
++--------------------------------------------------------+
+1 row in set (0.01 sec)
+
+mysql> select tidb_version()\G
+*************************** 1. row ***************************
+tidb_version(): Release Version: v3.0.0-beta-321-gc6a757b24
+Git Commit Hash: c6a757b2435965a72ada9d61ea8b7a367bd89b5f
+Git Branch: master
+UTC Build Time: 2019-04-02 02:38:04
+GoVersion: go version go1.12.1 linux/amd64
+Race Enabled: false
+TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e
+Check Table Before Drop: false
+1 row in set (0.00 sec)
+
+mysql>
+```",1,json key is not same with mysql bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error sql select json keys null select json keys null b select json keys null null select json keys a b e foo b a select json keys a b e foo b null what did you expect to see in mysql sql mysql select version version row in set sec mysql select json keys null json keys null null row in set sec mysql select json keys null b json keys null b null row in set sec mysql select json keys null null json keys null null null row in set sec mysql select json keys a b e foo b a json keys a b e foo b a null row in set sec mysql select json keys a b e foo b null json keys a b e foo b null null row in set sec mysql what did you see instead in tidb sql mysql select json keys null json keys null null row in set sec mysql select json keys null b json keys null b null row in set sec mysql select json keys null null json keys null null null row in set sec mysql select json keys a b e foo b a error invalid data type for json data mysql select json keys a b e foo b null json keys a b e foo b null null row in set sec mysql select tidb version g row tidb version release version beta git commit hash git branch master utc build time goversion go version linux race enabled false tikv min version alpha check table before drop false row in set sec mysql ,1
+13869,16593796542.0,IssuesEvent,2021-06-01 10:59:39,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,Support column type conversion between non-varchar to varchar,compatibility-breaker sig/sql-infra type/feature-request,"## Development Task
+
+We forbid changing column type conversion between non-varchar to varchar in master/v5.0.x(PR: #23688). Before #23688, this feature(conversion between char to varchar) was originally supported by modifying only the column structure without modifying the data.
+However, it was later found that some scenarios(these cases are executed on the branch which before forbidding this feature) could not be handled in this way, such as:
+```
+tidb> create table t(a int, b varchar(10), primary key(a, b));
+Query OK, 0 rows affected (0.01 sec)
+tidb> insert into t values (1, ""aaa "");
+Query OK, 1 row affected (0.00 sec)
+tidb> alter table t change column b b char(10);
+Query OK, 0 rows affected (0.01 sec)
+tidb> Create table t1 like t;
+Query OK, 0 rows affected (0.00 sec)
+tidb> insert into t1 values (1, ""aaa "");
+Query OK, 1 row affected (0.01 sec)
+
+tidb> select * from t;
++---+--------+
+| a | b |
++---+--------+
+| 1 | aaa |
++---+--------+
+1 row in set (0.00 sec)
+tidb> select * from t1;
++---+-----+
+| a | b |
++---+-----+
+| 1 | aaa |
++---+-----+
+1 row in set (0.00 sec)
+
+tidb> admin check table t;
+Query OK, 0 rows affected (0.01 sec)
+```
+
+Currently, we are working to support changing column types (not only changing column properties but also changing how column data is handled). This feature can also be supported by the way to modify the data which need to set`SET GLOBAL TIDB_ENABLE_CHANGE_COLUMN_TYPE = 1`.
+
+
+## Materials
+[old way to column type change](https://github.com/pingcap/tidb/blob/master/docs/design/2018-10-08-online-DDL.md#modify-column-operation)
+[new way to to column type change proposal](https://github.com/pingcap/tidb/blob/master/docs/design/2020-07-07-change-column-types.md)
+Related to issue: #23624, #23852.",True,"Support column type conversion between non-varchar to varchar - ## Development Task
+
+We forbid changing column type conversion between non-varchar to varchar in master/v5.0.x(PR: #23688). Before #23688, this feature(conversion between char to varchar) was originally supported by modifying only the column structure without modifying the data.
+However, it was later found that some scenarios(these cases are executed on the branch which before forbidding this feature) could not be handled in this way, such as:
+```
+tidb> create table t(a int, b varchar(10), primary key(a, b));
+Query OK, 0 rows affected (0.01 sec)
+tidb> insert into t values (1, ""aaa "");
+Query OK, 1 row affected (0.00 sec)
+tidb> alter table t change column b b char(10);
+Query OK, 0 rows affected (0.01 sec)
+tidb> Create table t1 like t;
+Query OK, 0 rows affected (0.00 sec)
+tidb> insert into t1 values (1, ""aaa "");
+Query OK, 1 row affected (0.01 sec)
+
+tidb> select * from t;
++---+--------+
+| a | b |
++---+--------+
+| 1 | aaa |
++---+--------+
+1 row in set (0.00 sec)
+tidb> select * from t1;
++---+-----+
+| a | b |
++---+-----+
+| 1 | aaa |
++---+-----+
+1 row in set (0.00 sec)
+
+tidb> admin check table t;
+Query OK, 0 rows affected (0.01 sec)
+```
+
+Currently, we are working to support changing column types (not only changing column properties but also changing how column data is handled). This feature can also be supported by the way to modify the data which need to set`SET GLOBAL TIDB_ENABLE_CHANGE_COLUMN_TYPE = 1`.
+
+
+## Materials
+[old way to column type change](https://github.com/pingcap/tidb/blob/master/docs/design/2018-10-08-online-DDL.md#modify-column-operation)
+[new way to to column type change proposal](https://github.com/pingcap/tidb/blob/master/docs/design/2020-07-07-change-column-types.md)
+Related to issue: #23624, #23852.",1,support column type conversion between non varchar to varchar development task we forbid changing column type conversion between non varchar to varchar in master x pr before this feature conversion between char to varchar was originally supported by modifying only the column structure without modifying the data however it was later found that some scenarios these cases are executed on the branch which before forbidding this feature could not be handled in this way such as tidb create table t a int b varchar primary key a b query ok rows affected sec tidb insert into t values aaa query ok row affected sec tidb alter table t change column b b char query ok rows affected sec tidb create table like t query ok rows affected sec tidb insert into values aaa query ok row affected sec tidb select from t a b aaa row in set sec tidb select from a b aaa row in set sec tidb admin check table t query ok rows affected sec currently we are working to support changing column types not only changing column properties but also changing how column data is handled this feature can also be supported by the way to modify the data which need to set set global tidb enable change column type materials related to issue ,1
+7329,9575840046.0,IssuesEvent,2019-05-07 07:37:23,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,addition between `time` and `interval` is not compatible with Mysql,for new contributors help wanted type/bug type/compatibility,"## Bug Report
+
+Please answer these questions before submitting your issue. Thanks!
+
+1. What did you do?
+If possible, provide a recipe for reproducing the error.
+
+```sql
+SELECT time('12:34:56') + INTERVAL 1 SECOND;
+```
+
+2. What did you expect to see?
+In Mysql
+```
+mysql> select time('12:34:56') + interval 1 second;
++--------------------------------------+
+| time('12:34:56') + interval 1 second |
++--------------------------------------+
+| 12:34:57 |
++--------------------------------------+
+1 row in set (0.00 sec)
+```
+
+
+3. What did you see instead?
+In TiDB
+```
+mysql> select time('12:34:56') + interval 1 second;
++--------------------------------------+
+| time('12:34:56') + interval 1 second |
++--------------------------------------+
+| 2019-04-30 12:34:57 |
++--------------------------------------+
+1 row in set (0.00 sec)
+```
+
+
+
+4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
+```tidb_version(): Release Version: v3.0.0-beta-463-g36bb1ae89```
+
+
+In mysql's document:
+
+`time('12:34:56') + interval 1 second` is same as `date_add(time('12:34:56'), interval 1 second)`
+
+And
+
+> The date argument specifies the starting date or datetime value. expr is an expression specifying the interval value to be added or subtracted from the starting date.",True,"addition between `time` and `interval` is not compatible with Mysql - ## Bug Report
+
+Please answer these questions before submitting your issue. Thanks!
+
+1. What did you do?
+If possible, provide a recipe for reproducing the error.
+
+```sql
+SELECT time('12:34:56') + INTERVAL 1 SECOND;
+```
+
+2. What did you expect to see?
+In Mysql
+```
+mysql> select time('12:34:56') + interval 1 second;
++--------------------------------------+
+| time('12:34:56') + interval 1 second |
++--------------------------------------+
+| 12:34:57 |
++--------------------------------------+
+1 row in set (0.00 sec)
+```
+
+
+3. What did you see instead?
+In TiDB
+```
+mysql> select time('12:34:56') + interval 1 second;
++--------------------------------------+
+| time('12:34:56') + interval 1 second |
++--------------------------------------+
+| 2019-04-30 12:34:57 |
++--------------------------------------+
+1 row in set (0.00 sec)
+```
+
+
+
+4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
+```tidb_version(): Release Version: v3.0.0-beta-463-g36bb1ae89```
+
+
+In mysql's document:
+
+`time('12:34:56') + interval 1 second` is same as `date_add(time('12:34:56'), interval 1 second)`
+
+And
+
+> The date argument specifies the starting date or datetime value. expr is an expression specifying the interval value to be added or subtracted from the starting date.",1,addition between time and interval is not compatible with mysql bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error sql select time interval second what did you expect to see in mysql mysql select time interval second time interval second row in set sec what did you see instead in tidb mysql select time interval second time interval second row in set sec what version of tidb are you using tidb server v or run select tidb version on tidb tidb version release version beta in mysql s document time interval second is same as date add time interval second and the date argument specifies the starting date or datetime value expr is an expression specifying the interval value to be added or subtracted from the starting date ,1
+611074,18944249431.0,IssuesEvent,2021-11-18 08:24:55,metabase/metabase,https://api.github.com/repos/metabase/metabase,closed,Allow to make scatter plots with categorical feature for y axis,Priority:P3 Type:New Feature Visualization/Charts,"Unfortunately when you want to use the metabase scatter plot for visualization it only accepts numerical data for y axis.
+
+Suppose I want a 2-d plot which x and y axis are my two categorical features and the size of bubble for the number of occurrences.
+
+It is a very common plot but metabase does not support it. You can choose your x axis from one of your categorical columns but y axis must be numerical which is not the case in most of scatter plot usage scenarios.
+
+:arrow_down: Please click the :+1: reaction instead of leaving a `+1` or `update?` comment",1.0,"Allow to make scatter plots with categorical feature for y axis - Unfortunately when you want to use the metabase scatter plot for visualization it only accepts numerical data for y axis.
+
+Suppose I want a 2-d plot which x and y axis are my two categorical features and the size of bubble for the number of occurrences.
+
+It is a very common plot but metabase does not support it. You can choose your x axis from one of your categorical columns but y axis must be numerical which is not the case in most of scatter plot usage scenarios.
+
+:arrow_down: Please click the :+1: reaction instead of leaving a `+1` or `update?` comment",0,allow to make scatter plots with categorical feature for y axis unfortunately when you want to use the metabase scatter plot for visualization it only accepts numerical data for y axis suppose i want a d plot which x and y axis are my two categorical features and the size of bubble for the number of occurrences it is a very common plot but metabase does not support it you can choose your x axis from one of your categorical columns but y axis must be numerical which is not the case in most of scatter plot usage scenarios arrow down please click the reaction instead of leaving a or update comment,0
+9812,8180821244.0,IssuesEvent,2018-08-28 20:42:25,grails/grails-core,https://api.github.com/repos/grails/grails-core,closed,Certificate at https://repo.grails.org has expired ,In Progress Infrastructure,"The certificate at https://repo.grails.org has expired and is failing our builds.
+
+",1.0,"Certificate at https://repo.grails.org has expired - The certificate at https://repo.grails.org has expired and is failing our builds.
+
+",0,certificate at has expired the certificate at has expired and is failing our builds img width alt screen shot at pm src ,0
+8080,10208742730.0,IssuesEvent,2019-08-14 10:54:16,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,JDBC to TiDB useCursorFetch error...,component/mysql-protocol type/bug type/compatibility,"## Bug Report
+
+Please answer these questions before submitting your issue. Thanks!
+
+1. What did you do?
+If possible, provide a recipe for reproducing the error.
+
+the jdbc url is :
+jdbc:mysql://xxx?serverTimezone=Asia/Shanghai&max_allowed_packet=-1&autoReconnect=true&useUnicode=true&characterEncoding=utf8&useSSL=true&useCursorFetch=true&defaultFetchSize=10000
+
+
+org.springframework.jdbc.UncategorizedSQLException: StatementCallback; uncategorized SQLException for SQL [select max(marstime) from xxx.Account]; SQL state [HY000]; error code [1105]; ERROR 1105 (HY000): unsupported flag 1; nested exception is java.sql.SQLException: ERROR 1105 (HY000): unsupported flag 1
+ at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:89)
+ at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:81)
+ at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:81)
+ at org.springframework.jdbc.core.JdbcTemplate.translateException(JdbcTemplate.java:1414)
+ at org.springframework.jdbc.core.JdbcTemplate.execute(JdbcTemplate.java:388)
+ at org.springframework.jdbc.core.JdbcTemplate.query(JdbcTemplate.java:452)
+ at org.springframework.jdbc.core.JdbcTemplate.query(JdbcTemplate.java:462)
+ at org.springframework.jdbc.core.JdbcTemplate.queryForObject(JdbcTemplate.java:473)
+ at org.springframework.jdbc.core.JdbcTemplate.queryForObject(JdbcTemplate.java:480)
+ at com.netease.mars.source.AccountSource.run(AccountSource.java:106)
+ at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:94)
+ at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:58)
+ at org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:99)
+ at org.apache.flink.streaming.runtime.tasks.StoppableSourceStreamTask.run(StoppableSourceStreamTask.java:45)
+ at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:300)
+ at org.apache.flink.runtime.taskmanager.Task.run(Task.java:704)
+ at java.lang.Thread.run(Thread.java:748)
+Caused by: java.sql.SQLException: ERROR 1105 (HY000): unsupported flag 1
+ at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129)
+ at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:97)
+ at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122)
+ at com.mysql.cj.jdbc.ServerPreparedStatement.serverExecute(ServerPreparedStatement.java:633)
+ at com.mysql.cj.jdbc.ServerPreparedStatement.executeInternal(ServerPreparedStatement.java:414)
+ at com.mysql.cj.jdbc.ClientPreparedStatement.execute(ClientPreparedStatement.java:392)
+ at com.mysql.cj.jdbc.StatementImpl.createResultSetUsingServerFetch(StatementImpl.java:582)
+ at com.mysql.cj.jdbc.StatementImpl.executeQuery(StatementImpl.java:1158)
+ at com.alibaba.druid.pool.DruidPooledStatement.executeQuery(DruidPooledStatement.java:230)
+ at org.springframework.jdbc.core.JdbcTemplate$1QueryStatementCallback.doInStatement(JdbcTemplate.java:439)
+ at org.springframework.jdbc.core.JdbcTemplate.execute(JdbcTemplate.java:376)
+
+
+2. What did you expect to see?
+
+
+
+3. What did you see instead?
+
+
+
+4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
+
+
+mysql> select tidb_version();
++--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+| tidb_version() |
++--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+| Release Version: v2.0.11
+Git Commit Hash: 83889a5d821f40412be1af2ad2f2393737f517fd
+Git Branch: release-2.0
+UTC Build Time: 2019-01-03 09:27:12
+GoVersion: go version go1.11.2 linux/amd64
+TiKV Min Version: 2.0.0-rc.4.1
+",True,"JDBC to TiDB useCursorFetch error... - ## Bug Report
+
+Please answer these questions before submitting your issue. Thanks!
+
+1. What did you do?
+If possible, provide a recipe for reproducing the error.
+
+the jdbc url is :
+jdbc:mysql://xxx?serverTimezone=Asia/Shanghai&max_allowed_packet=-1&autoReconnect=true&useUnicode=true&characterEncoding=utf8&useSSL=true&useCursorFetch=true&defaultFetchSize=10000
+
+
+org.springframework.jdbc.UncategorizedSQLException: StatementCallback; uncategorized SQLException for SQL [select max(marstime) from xxx.Account]; SQL state [HY000]; error code [1105]; ERROR 1105 (HY000): unsupported flag 1; nested exception is java.sql.SQLException: ERROR 1105 (HY000): unsupported flag 1
+ at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:89)
+ at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:81)
+ at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:81)
+ at org.springframework.jdbc.core.JdbcTemplate.translateException(JdbcTemplate.java:1414)
+ at org.springframework.jdbc.core.JdbcTemplate.execute(JdbcTemplate.java:388)
+ at org.springframework.jdbc.core.JdbcTemplate.query(JdbcTemplate.java:452)
+ at org.springframework.jdbc.core.JdbcTemplate.query(JdbcTemplate.java:462)
+ at org.springframework.jdbc.core.JdbcTemplate.queryForObject(JdbcTemplate.java:473)
+ at org.springframework.jdbc.core.JdbcTemplate.queryForObject(JdbcTemplate.java:480)
+ at com.netease.mars.source.AccountSource.run(AccountSource.java:106)
+ at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:94)
+ at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:58)
+ at org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:99)
+ at org.apache.flink.streaming.runtime.tasks.StoppableSourceStreamTask.run(StoppableSourceStreamTask.java:45)
+ at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:300)
+ at org.apache.flink.runtime.taskmanager.Task.run(Task.java:704)
+ at java.lang.Thread.run(Thread.java:748)
+Caused by: java.sql.SQLException: ERROR 1105 (HY000): unsupported flag 1
+ at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129)
+ at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:97)
+ at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122)
+ at com.mysql.cj.jdbc.ServerPreparedStatement.serverExecute(ServerPreparedStatement.java:633)
+ at com.mysql.cj.jdbc.ServerPreparedStatement.executeInternal(ServerPreparedStatement.java:414)
+ at com.mysql.cj.jdbc.ClientPreparedStatement.execute(ClientPreparedStatement.java:392)
+ at com.mysql.cj.jdbc.StatementImpl.createResultSetUsingServerFetch(StatementImpl.java:582)
+ at com.mysql.cj.jdbc.StatementImpl.executeQuery(StatementImpl.java:1158)
+ at com.alibaba.druid.pool.DruidPooledStatement.executeQuery(DruidPooledStatement.java:230)
+ at org.springframework.jdbc.core.JdbcTemplate$1QueryStatementCallback.doInStatement(JdbcTemplate.java:439)
+ at org.springframework.jdbc.core.JdbcTemplate.execute(JdbcTemplate.java:376)
+
+
+2. What did you expect to see?
+
+
+
+3. What did you see instead?
+
+
+
+4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
+
+
+mysql> select tidb_version();
++--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+| tidb_version() |
++--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
+| Release Version: v2.0.11
+Git Commit Hash: 83889a5d821f40412be1af2ad2f2393737f517fd
+Git Branch: release-2.0
+UTC Build Time: 2019-01-03 09:27:12
+GoVersion: go version go1.11.2 linux/amd64
+TiKV Min Version: 2.0.0-rc.4.1
+",1,jdbc to tidb usecursorfetch error bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error the jdbc url is : jdbc mysql xxx servertimezone asia shanghai max allowed packet autoreconnect true useunicode true characterencoding usessl true usecursorfetch true defaultfetchsize org springframework jdbc uncategorizedsqlexception statementcallback uncategorized sqlexception for sql sql state error code error unsupported flag nested exception is java sql sqlexception error unsupported flag at org springframework jdbc support abstractfallbacksqlexceptiontranslator translate abstractfallbacksqlexceptiontranslator java at org springframework jdbc support abstractfallbacksqlexceptiontranslator translate abstractfallbacksqlexceptiontranslator java at org springframework jdbc support abstractfallbacksqlexceptiontranslator translate abstractfallbacksqlexceptiontranslator java at org springframework jdbc core jdbctemplate translateexception jdbctemplate java at org springframework jdbc core jdbctemplate execute jdbctemplate java at org springframework jdbc core jdbctemplate query jdbctemplate java at org springframework jdbc core jdbctemplate query jdbctemplate java at org springframework jdbc core jdbctemplate queryforobject jdbctemplate java at org springframework jdbc core jdbctemplate queryforobject jdbctemplate java at com netease mars source accountsource run accountsource java at org apache flink streaming api operators streamsource run streamsource java at org apache flink streaming api operators streamsource run streamsource java at org apache flink streaming runtime tasks sourcestreamtask run sourcestreamtask java at org apache flink streaming runtime tasks stoppablesourcestreamtask run stoppablesourcestreamtask java at org apache flink streaming runtime tasks streamtask invoke streamtask java at org apache flink runtime taskmanager task run task java at java lang thread run thread java caused by java sql sqlexception error unsupported flag at com mysql cj jdbc exceptions sqlerror createsqlexception sqlerror java at com mysql cj jdbc exceptions sqlerror createsqlexception sqlerror java at com mysql cj jdbc exceptions sqlexceptionsmapping translateexception sqlexceptionsmapping java at com mysql cj jdbc serverpreparedstatement serverexecute serverpreparedstatement java at com mysql cj jdbc serverpreparedstatement executeinternal serverpreparedstatement java at com mysql cj jdbc clientpreparedstatement execute clientpreparedstatement java at com mysql cj jdbc statementimpl createresultsetusingserverfetch statementimpl java at com mysql cj jdbc statementimpl executequery statementimpl java at com alibaba druid pool druidpooledstatement executequery druidpooledstatement java at org springframework jdbc core jdbctemplate doinstatement jdbctemplate java at org springframework jdbc core jdbctemplate execute jdbctemplate java what did you expect to see what did you see instead what version of tidb are you using tidb server v or run select tidb version on tidb mysql select tidb version tidb version release version git commit hash git branch release utc build time goversion go version linux tikv min version rc ,1
+17606,3012747759.0,IssuesEvent,2015-07-29 02:09:39,yawlfoundation/yawl,https://api.github.com/repos/yawlfoundation/yawl,closed,[CLOSED] Resourcing information to be checked by editor,auto-migrated Milestone-Release2.0 Priority-High Type-Defect,"
**Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
+_Monday Jul 27, 2015 at 03:20 GMT_
+_Originally opened as https://github.com/adamsmj/yawl/issues/9_
-
Type: Upgrade version
-Release Date: 2018-05-08
-Fix Resolution: Microsoft.AspNetCore.Server.Kestrel.Core - 2.0.3,2.1.0;Microsoft.AspNetCore.Server.Kestrel.Transport.Abstractions - 2.0.3,2.1.0;Microsoft.AspNetCore.Server.Kestrel.Transport.Libuv - 2.0.3,2.1.0;Microsoft.AspNetCore.All - 2.0.8,2.1.0
+---- - +``` +The editor should check whether the resourcing information that is +specified for a certain specification is still in line with the latest in +the resourcing database (e.g. suppose a role/participant is used that has +been deleted in the mean time). - + +``` -+ +``` +Original issue reported on code.google.com by `arthurte...@gmail.com` on 18 Jul 2008 at 4:52 +",0, resourcing information to be checked by editor issue by monday jul at gmt originally opened as the editor should check whether the resourcing information that is specified for a certain specification is still in line with the latest in the resourcing database e g suppose a role participant is used that has been deleted in the mean time original issue reported on code google com by arthurte gmail com on jul at ,0 +305394,26387066255.0,IssuesEvent,2023-01-12 13:06:52,rhinstaller/kickstart-tests,https://api.github.com/repos/rhinstaller/kickstart-tests,closed,repo-include failing on rhel8 and rhel9: *** wrong version of package-1 was installed,disabled test,"repo-include test is failing on rhel8 and rhel9 scenarios: +``` +stdout: +*** wrong version of package-1 was installed +*** unable to find marker for package-1 + +2022-02-17 01:22:42,172 INFO: RESULT:repo-include:7f8a0daed3b6:FAILED:Validation failed with return code 1 +``` +[kstest.log](https://github.com/rhinstaller/kickstart-tests/files/8086557/kstest.log) +[virt-install.log](https://github.com/rhinstaller/kickstart-tests/files/8086559/virt-install.log) +",1.0,"repo-include failing on rhel8 and rhel9: *** wrong version of package-1 was installed - repo-include test is failing on rhel8 and rhel9 scenarios: +``` +stdout: +*** wrong version of package-1 was installed +*** unable to find marker for package-1 + +2022-02-17 01:22:42,172 INFO: RESULT:repo-include:7f8a0daed3b6:FAILED:Validation failed with return code 1 +``` +[kstest.log](https://github.com/rhinstaller/kickstart-tests/files/8086557/kstest.log) +[virt-install.log](https://github.com/rhinstaller/kickstart-tests/files/8086559/virt-install.log) +",0,repo include failing on and wrong version of package was installed repo include test is failing on and scenarios stdout wrong version of package was installed unable to find marker for package info result repo include failed validation failed with return code ,0 +558234,16528944914.0,IssuesEvent,2021-05-27 01:31:47,eclipse-ee4j/glassfish,https://api.github.com/repos/eclipse-ee4j/glassfish,closed,Provide default value for all attributes of the config bean without requiring the instance be created first.,Component: rest-interface ERR: Assignee Priority: Major Stale Type: New Feature,"This is related to #20192. +REST should be able to provide the default value just by given the ConfigBean type, eg. jdbc-resource, thread-pool, virtual-server etc. +But currently unless this attribute is one of the create command param, it cannot do that. +The limitation is that not all default value is available unless the instance is created first. -### microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg
+Want to request REST to provide a way to give all the default value of any attributes without requiring: +1\. the attribute be one of the params for create command +2\. the resource has to be created first to be able to get those default value. +#### Affected Versions +[4.0_dev]",1.0,"Provide default value for all attributes of the config bean without requiring the instance be created first. - This is related to #20192. +REST should be able to provide the default value just by given the ConfigBean type, eg. jdbc-resource, thread-pool, virtual-server etc. +But currently unless this attribute is one of the create command param, it cannot do that. +The limitation is that not all default value is available unless the instance is created first. -ASP.NET Core components for working with the IIS AspNetCoreModule.
-Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg
-Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
-Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.iisintegration/2.0.0/microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg
-+Want to request REST to provide a way to give all the default value of any attributes without requiring: +1\. the attribute be one of the params for create command +2\. the resource has to be created first to be able to get those default value. +#### Affected Versions +[4.0_dev]",0,provide default value for all attributes of the config bean without requiring the instance be created first this is related to rest should be able to provide the default value just by given the configbean type eg jdbc resource thread pool virtual server etc but currently unless this attribute is one of the create command param it cannot do that the limitation is that not all default value is available unless the instance is created first want to request rest to provide a way to give all the default value of any attributes without requiring the attribute be one of the params for create command the resource has to be created first to be able to get those default value affected versions ,0 +203256,15359239791.0,IssuesEvent,2021-03-01 15:38:44,hashgraph/hedera-mirror-node,https://api.github.com/repos/hashgraph/hedera-mirror-node,opened,RecordFileParserPerformanceTest times out in CircleCi,P3 bug test," + +**Detailed Description** +https://app.circleci.com/pipelines/github/hashgraph/hedera-mirror-node/6248/workflows/c52d402b-7bf3-4b49-b386-91ad18116820/jobs/17965/steps + +``` +2021-03-01T09:23:28.771-0600 INFO main c.h.m.i.p.r.RecordFileParserPerformanceTest No active profile set, falling back to default profiles: default +2021-03-01T09:23:32.694-0600 INFO main c.h.m.i.c.MetricsConfiguration Collecting table metrics: [] +2021-03-01T09:23:32.738-0600 INFO main c.h.m.i.c.CredentialsProviderConfiguration Setting up S3 async client using AWS Default Credentials Provider +2021-03-01T09:23:33.044-0600 INFO main o.f.c.i.c.DbMigrate Current version of schema ""public"": << Empty Schema >> +2021-03-01T09:23:33.077-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.0 - Init"" +2021-03-01T09:23:33.207-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.1 - update"" +2021-03-01T09:23:33.228-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.2 - procedures"" +2021-03-01T09:23:33.257-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.3 - procedures"" +2021-03-01T09:23:33.268-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.4 - update"" +2021-03-01T09:23:33.280-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.5 - file create patch"" +2021-03-01T09:23:33.289-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.7 - application status"" +2021-03-01T09:23:33.301-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.9 - index"" +2021-03-01T09:23:33.308-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.9.1 - transactions indexes"" +2021-03-01T09:23:33.317-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.10.0 - account balances"" +2021-03-01T09:23:33.336-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.10.1 - freeze transaction"" +2021-03-01T09:23:33.346-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.10.2 - response codes"" +2021-03-01T09:23:33.353-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.10.3 - account balances"" +2021-03-01T09:23:33.364-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.0 - entity key"" +2021-03-01T09:23:33.374-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.1 - remove deprecated"" +2021-03-01T09:23:33.389-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.2 - transactions pk"" +2021-03-01T09:23:33.420-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.3 - entity key"" +2021-03-01T09:23:33.425-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.4 - duplicate entities"" +2021-03-01T09:23:33.434-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.5 - entity create"" +2021-03-01T09:23:33.440-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.5.1 - response codes"" +2021-03-01T09:23:33.447-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.6 - Missing Entities"" +2021-03-01T09:23:33.449-0600 WARN main c.h.m.i.m.V1_11_6__Missing_Entities Skipping entity import due to missing file /root/project/hedera-mirror-importer/./data/accountInfo.txt +2021-03-01T09:23:33.455-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.12 - transactions maxfee duration"" +2021-03-01T09:23:33.459-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.13 - transaction hash"" +2021-03-01T09:23:33.464-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.14 - response codes"" +2021-03-01T09:23:33.469-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.15 - unknown transaction types"" +2021-03-01T09:23:33.483-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.16 - cryptotransferlist realm num"" +2021-03-01T09:23:33.496-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.0 - hcs support"" +2021-03-01T09:23:33.509-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.1 - f entity create"" +2021-03-01T09:23:33.517-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.2 - transaction bytes"" +2021-03-01T09:23:33.522-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.3 - drop f entity create"" +2021-03-01T09:23:33.528-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.4 - hcs proto changes"" +2021-03-01T09:23:33.538-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.5 - fix mirror api permissions"" +2021-03-01T09:23:33.545-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.18.0 - non fee transfers"" +2021-03-01T09:23:33.554-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.19.0 - drop fks trigger unused indexes"" +2021-03-01T09:23:33.568-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.20 - drop file id"" +2021-03-01T09:23:33.575-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.21 - add topicmessage seqindex"" +2021-03-01T09:23:33.581-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.22 - remove unused components"" +2021-03-01T09:23:33.590-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.23.0 - r5 rc4 proto updates"" +2021-03-01T09:23:33.596-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.23.1 - add topicmessage runninghashversion"" +2021-03-01T09:23:33.601-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.24.0 - remove bypass state"" +2021-03-01T09:23:33.606-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.24.1 - t record files timestamps"" +2021-03-01T09:23:33.616-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.25.0 - denormalize entities"" +2021-03-01T09:23:33.633-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.25.1 - denormalize transfer tables"" +2021-03-01T09:23:33.643-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.26.0 - topic message fragmentation"" +2021-03-01T09:23:33.648-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.26.1 - pg notify"" +2021-03-01T09:23:33.653-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.26.2 - drop notify trigger"" +2021-03-01T09:23:33.657-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.27.0 - add event streams"" +2021-03-01T09:23:33.663-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.27.1 - populate consensus start end for record files"" +2021-03-01T09:23:33.674-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.27.2 - update name of record file to filename"" +2021-03-01T09:23:33.679-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.27.3 - rename tables"" +2021-03-01T09:23:33.685-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.28.0 - address book"" +2021-03-01T09:23:33.697-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.28.2 - record file index consensus end"" +2021-03-01T09:23:33.703-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.29.0 - store verification node in stream file tables"" +2021-03-01T09:23:33.718-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.30.0 - balance entity id"" +2021-03-01T09:23:33.728-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.30.1 - hts support"" +2021-03-01T09:23:33.744-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.31.0 - fix address book 102"" +2021-03-01T09:23:33.749-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.31.1 - token symbol size 100"" +2021-03-01T09:23:33.754-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.31.2 - remove invalid entities"" +2021-03-01T09:23:33.768-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.31.3 - delete token transfers transaction type"" +2021-03-01T09:23:33.772-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.32.1 - update transaction index"" +2021-03-01T09:23:33.778-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.33.0 - drop token account id"" +2021-03-01T09:23:33.791-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.33.1 - update record file"" +2021-03-01T09:23:33.808-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.34.0 - schedule support"" +2021-03-01T09:23:33.819-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.34.1 - transaction schedule"" +2021-03-01T09:23:33.830-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.34.2 - rename record file hash"" +2021-03-01T09:23:33.838-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.35.0 - update schedule response codes"" +2021-03-01T09:23:33.846-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.35.1 - remove filesystem"" +2021-03-01T09:23:33.859-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.35.2 - schedule index"" +2021-03-01T09:23:33.866-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.35.3 - record file block index"" +2021-03-01T09:23:33.875-0600 INFO main o.f.c.i.c.DbMigrate Successfully applied 69 migrations to schema ""public"" (execution time 00:00.843s) +2021-03-01T09:23:36.989-0600 INFO main c.h.m.i.c.MirrorImporterConfiguration Configured to download from S3 in region us-east-1 with bucket name 'test' +2021-03-01T09:23:41.482-0600 INFO main c.h.m.i.p.r.RecordFileParserPerformanceTest Started RecordFileParserPerformanceTest in 14.229 seconds (JVM running for 17.578) +2021-03-01T09:23:42.404-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_02_56.025429000Z.rcd +2021-03-01T09:23:43.013-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_02_58.019176000Z.rcd +2021-03-01T09:23:43.236-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_00.032295000Z.rcd +2021-03-01T09:23:43.488-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_02.031984000Z.rcd +2021-03-01T09:23:43.851-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_04.004057000Z.rcd +2021-03-01T09:23:56.260-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_06.001417000Z.rcd +2021-03-01T09:23:56.362-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_08.002071000Z.rcd +2021-03-01T09:23:56.497-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_10.016188000Z.rcd +2021-03-01T09:23:56.630-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_12.001771000Z.rcd +2021-03-01T09:23:56.988-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_14.015461000Z.rcd +2021-03-01T09:23:57.077-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_16.061108000Z.rcd +2021-03-01T09:23:57.992-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_18.004362000Z.rcd +2021-03-01T09:23:58.407-0600 INFO main c.h.m.i.p.r.RecordFileParserPerformanceTest Executing: parse() +2021-03-01T09:23:58.556-0600 INFO main c.h.m.i.c.MirrorDateRangePropertiesProcessor RECORD: parser will parse items in the range [1970-01-01T00:00:00Z, 2262-04-11T23:47:16.854775807Z] +2021-03-01T09:24:00.892-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 12654 rows to transaction table in 343.9 ms +2021-03-01T09:24:01.439-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 50588 rows to crypto_transfer table in 483.9 ms +2021-03-01T09:24:01.516-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 17 entities in 76.62 ms +2021-03-01T09:24:01.516-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 2.097 s +2021-03-01T09:24:01.573-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 12654 transactions from record file 2021-01-14T20_02_56.025429000Z.rcd in 3.118 s. Success: true +2021-03-01T09:24:02.475-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 12378 rows to transaction table in 546.3 ms +2021-03-01T09:24:03.535-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 49477 rows to crypto_transfer table in 1.039 s +2021-03-01T09:24:03.548-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 17 entities in 12.97 ms +2021-03-01T09:24:03.548-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.786 s +2021-03-01T09:24:03.550-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 12378 transactions from record file 2021-01-14T20_02_58.019176000Z.rcd in 1.892 s. Success: true +2021-03-01T09:24:04.588-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11649 rows to transaction table in 488.3 ms +2021-03-01T09:24:04.757-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 46575 rows to crypto_transfer table in 149.5 ms +2021-03-01T09:24:04.770-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 17 entities in 12.78 ms +2021-03-01T09:24:04.770-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.117 s +2021-03-01T09:24:04.771-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11649 transactions from record file 2021-01-14T20_03_00.032295000Z.rcd in 1.184 s. Success: true +2021-03-01T09:24:06.210-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11196 rows to transaction table in 1.325 s +2021-03-01T09:24:06.425-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 44753 rows to crypto_transfer table in 190.2 ms +2021-03-01T09:24:06.437-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 11.89 ms +2021-03-01T09:24:06.437-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.592 s +2021-03-01T09:24:06.439-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11196 transactions from record file 2021-01-14T20_03_02.031984000Z.rcd in 1.650 s. Success: true +2021-03-01T09:24:07.169-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 10500 rows to transaction table in 626.7 ms +2021-03-01T09:24:07.906-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 41979 rows to crypto_transfer table in 719.5 ms +2021-03-01T09:24:07.919-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 12.65 ms +2021-03-01T09:24:07.919-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.422 s +2021-03-01T09:24:07.921-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 10500 transactions from record file 2021-01-14T20_03_04.004057000Z.rcd in 1.473 s. Success: true +2021-03-01T09:24:08.607-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11427 rows to transaction table in 551.0 ms +2021-03-01T09:24:09.021-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 45665 rows to crypto_transfer table in 396.4 ms +2021-03-01T09:24:09.033-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 17 entities in 12.04 ms +2021-03-01T09:24:09.034-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.047 s +2021-03-01T09:24:09.035-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11427 transactions from record file 2021-01-14T20_03_06.001417000Z.rcd in 1.107 s. Success: true +2021-03-01T09:24:09.424-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11771 rows to transaction table in 115.9 ms +2021-03-01T09:24:09.804-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 47053 rows to crypto_transfer table in 358.5 ms +2021-03-01T09:24:09.816-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 11.80 ms +2021-03-01T09:24:09.816-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 634.3 ms +2021-03-01T09:24:09.817-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11771 transactions from record file 2021-01-14T20_03_08.002071000Z.rcd in 719.5 ms. Success: true +2021-03-01T09:24:10.178-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 10895 rows to transaction table in 230.7 ms +2021-03-01T09:24:10.523-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 43554 rows to crypto_transfer table in 327.1 ms +2021-03-01T09:24:10.535-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 11.07 ms +2021-03-01T09:24:10.535-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 645.9 ms +2021-03-01T09:24:10.537-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 10895 transactions from record file 2021-01-14T20_03_10.016188000Z.rcd in 711.1 ms. Success: true +2021-03-01T09:24:10.739-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 10642 rows to transaction table in 98.91 ms +2021-03-01T09:24:10.930-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 42537 rows to crypto_transfer table in 170.9 ms +2021-03-01T09:24:10.943-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 13.00 ms +2021-03-01T09:24:10.944-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 341.4 ms +2021-03-01T09:24:10.947-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 10642 transactions from record file 2021-01-14T20_03_12.001771000Z.rcd in 400.9 ms. Success: true +2021-03-01T09:24:13.375-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11801 rows to transaction table in 2.272 s +2021-03-01T09:24:13.747-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 47174 rows to crypto_transfer table in 353.5 ms +2021-03-01T09:24:13.761-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 12.79 ms +2021-03-01T09:24:13.761-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 2.722 s +2021-03-01T09:24:13.762-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11801 transactions from record file 2021-01-14T20_03_14.015461000Z.rcd in 2.808 s. Success: true +2021-03-01T09:24:14.249-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11044 rows to transaction table in 302.3 ms +2021-03-01T09:24:14.583-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 44158 rows to crypto_transfer table in 321.3 ms +2021-03-01T09:24:14.602-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 10.98 ms +2021-03-01T09:24:14.602-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 694.9 ms +2021-03-01T09:24:14.603-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11044 transactions from record file 2021-01-14T20_03_16.061108000Z.rcd in 801.3 ms. Success: true +2021-03-01T09:24:14.910-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 10665 rows to transaction table in 226.0 ms +2021-03-01T09:24:15.265-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 42630 rows to crypto_transfer table in 338.9 ms +2021-03-01T09:24:15.276-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 10.03 ms +2021-03-01T09:24:15.276-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 618.4 ms +2021-03-01T09:24:15.277-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 10665 transactions from record file 2021-01-14T20_03_18.004362000Z.rcd in 663.4 ms. Success: true +[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 49.831 s <<< FAILURE! - in com.hedera.mirror.importer.parser.record.RecordFileParserPerformanceTest +[ERROR] com.hedera.mirror.importer.parser.record.RecordFileParserPerformanceTest.parse Time elapsed: 17.606 s <<< ERROR! +java.util.concurrent.TimeoutException: parse() timed out after 15 seconds +``` + +**Actual Behavior** +Steps to reproduce the behavior: +1. Go to '...' +2. Click on '....' +3. Scroll down to '....' + +**Expected Behavior** + +**Environment:** +- Java: [e.g. OpenJDK 11.0.4] +- Node: [e.g. v12.9.1-x86] +- OS: [e.g. Ubuntu 18.04] +- Version: [e.g. 1.0.1] + +**Additional Context** + +",1.0,"RecordFileParserPerformanceTest times out in CircleCi - + +**Detailed Description** +https://app.circleci.com/pipelines/github/hashgraph/hedera-mirror-node/6248/workflows/c52d402b-7bf3-4b49-b386-91ad18116820/jobs/17965/steps + +``` +2021-03-01T09:23:28.771-0600 INFO main c.h.m.i.p.r.RecordFileParserPerformanceTest No active profile set, falling back to default profiles: default +2021-03-01T09:23:32.694-0600 INFO main c.h.m.i.c.MetricsConfiguration Collecting table metrics: [] +2021-03-01T09:23:32.738-0600 INFO main c.h.m.i.c.CredentialsProviderConfiguration Setting up S3 async client using AWS Default Credentials Provider +2021-03-01T09:23:33.044-0600 INFO main o.f.c.i.c.DbMigrate Current version of schema ""public"": << Empty Schema >> +2021-03-01T09:23:33.077-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.0 - Init"" +2021-03-01T09:23:33.207-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.1 - update"" +2021-03-01T09:23:33.228-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.2 - procedures"" +2021-03-01T09:23:33.257-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.3 - procedures"" +2021-03-01T09:23:33.268-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.4 - update"" +2021-03-01T09:23:33.280-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.5 - file create patch"" +2021-03-01T09:23:33.289-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.7 - application status"" +2021-03-01T09:23:33.301-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.9 - index"" +2021-03-01T09:23:33.308-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.9.1 - transactions indexes"" +2021-03-01T09:23:33.317-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.10.0 - account balances"" +2021-03-01T09:23:33.336-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.10.1 - freeze transaction"" +2021-03-01T09:23:33.346-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.10.2 - response codes"" +2021-03-01T09:23:33.353-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.10.3 - account balances"" +2021-03-01T09:23:33.364-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.0 - entity key"" +2021-03-01T09:23:33.374-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.1 - remove deprecated"" +2021-03-01T09:23:33.389-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.2 - transactions pk"" +2021-03-01T09:23:33.420-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.3 - entity key"" +2021-03-01T09:23:33.425-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.4 - duplicate entities"" +2021-03-01T09:23:33.434-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.5 - entity create"" +2021-03-01T09:23:33.440-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.5.1 - response codes"" +2021-03-01T09:23:33.447-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.11.6 - Missing Entities"" +2021-03-01T09:23:33.449-0600 WARN main c.h.m.i.m.V1_11_6__Missing_Entities Skipping entity import due to missing file /root/project/hedera-mirror-importer/./data/accountInfo.txt +2021-03-01T09:23:33.455-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.12 - transactions maxfee duration"" +2021-03-01T09:23:33.459-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.13 - transaction hash"" +2021-03-01T09:23:33.464-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.14 - response codes"" +2021-03-01T09:23:33.469-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.15 - unknown transaction types"" +2021-03-01T09:23:33.483-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.16 - cryptotransferlist realm num"" +2021-03-01T09:23:33.496-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.0 - hcs support"" +2021-03-01T09:23:33.509-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.1 - f entity create"" +2021-03-01T09:23:33.517-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.2 - transaction bytes"" +2021-03-01T09:23:33.522-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.3 - drop f entity create"" +2021-03-01T09:23:33.528-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.4 - hcs proto changes"" +2021-03-01T09:23:33.538-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.17.5 - fix mirror api permissions"" +2021-03-01T09:23:33.545-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.18.0 - non fee transfers"" +2021-03-01T09:23:33.554-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.19.0 - drop fks trigger unused indexes"" +2021-03-01T09:23:33.568-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.20 - drop file id"" +2021-03-01T09:23:33.575-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.21 - add topicmessage seqindex"" +2021-03-01T09:23:33.581-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.22 - remove unused components"" +2021-03-01T09:23:33.590-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.23.0 - r5 rc4 proto updates"" +2021-03-01T09:23:33.596-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.23.1 - add topicmessage runninghashversion"" +2021-03-01T09:23:33.601-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.24.0 - remove bypass state"" +2021-03-01T09:23:33.606-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.24.1 - t record files timestamps"" +2021-03-01T09:23:33.616-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.25.0 - denormalize entities"" +2021-03-01T09:23:33.633-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.25.1 - denormalize transfer tables"" +2021-03-01T09:23:33.643-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.26.0 - topic message fragmentation"" +2021-03-01T09:23:33.648-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.26.1 - pg notify"" +2021-03-01T09:23:33.653-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.26.2 - drop notify trigger"" +2021-03-01T09:23:33.657-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.27.0 - add event streams"" +2021-03-01T09:23:33.663-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.27.1 - populate consensus start end for record files"" +2021-03-01T09:23:33.674-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.27.2 - update name of record file to filename"" +2021-03-01T09:23:33.679-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.27.3 - rename tables"" +2021-03-01T09:23:33.685-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.28.0 - address book"" +2021-03-01T09:23:33.697-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.28.2 - record file index consensus end"" +2021-03-01T09:23:33.703-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.29.0 - store verification node in stream file tables"" +2021-03-01T09:23:33.718-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.30.0 - balance entity id"" +2021-03-01T09:23:33.728-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.30.1 - hts support"" +2021-03-01T09:23:33.744-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.31.0 - fix address book 102"" +2021-03-01T09:23:33.749-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.31.1 - token symbol size 100"" +2021-03-01T09:23:33.754-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.31.2 - remove invalid entities"" +2021-03-01T09:23:33.768-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.31.3 - delete token transfers transaction type"" +2021-03-01T09:23:33.772-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.32.1 - update transaction index"" +2021-03-01T09:23:33.778-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.33.0 - drop token account id"" +2021-03-01T09:23:33.791-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.33.1 - update record file"" +2021-03-01T09:23:33.808-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.34.0 - schedule support"" +2021-03-01T09:23:33.819-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.34.1 - transaction schedule"" +2021-03-01T09:23:33.830-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.34.2 - rename record file hash"" +2021-03-01T09:23:33.838-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.35.0 - update schedule response codes"" +2021-03-01T09:23:33.846-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.35.1 - remove filesystem"" +2021-03-01T09:23:33.859-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.35.2 - schedule index"" +2021-03-01T09:23:33.866-0600 INFO main o.f.c.i.c.DbMigrate Migrating schema ""public"" to version ""1.35.3 - record file block index"" +2021-03-01T09:23:33.875-0600 INFO main o.f.c.i.c.DbMigrate Successfully applied 69 migrations to schema ""public"" (execution time 00:00.843s) +2021-03-01T09:23:36.989-0600 INFO main c.h.m.i.c.MirrorImporterConfiguration Configured to download from S3 in region us-east-1 with bucket name 'test' +2021-03-01T09:23:41.482-0600 INFO main c.h.m.i.p.r.RecordFileParserPerformanceTest Started RecordFileParserPerformanceTest in 14.229 seconds (JVM running for 17.578) +2021-03-01T09:23:42.404-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_02_56.025429000Z.rcd +2021-03-01T09:23:43.013-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_02_58.019176000Z.rcd +2021-03-01T09:23:43.236-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_00.032295000Z.rcd +2021-03-01T09:23:43.488-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_02.031984000Z.rcd +2021-03-01T09:23:43.851-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_04.004057000Z.rcd +2021-03-01T09:23:56.260-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_06.001417000Z.rcd +2021-03-01T09:23:56.362-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_08.002071000Z.rcd +2021-03-01T09:23:56.497-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_10.016188000Z.rcd +2021-03-01T09:23:56.630-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_12.001771000Z.rcd +2021-03-01T09:23:56.988-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_14.015461000Z.rcd +2021-03-01T09:23:57.077-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_16.061108000Z.rcd +2021-03-01T09:23:57.992-0600 INFO main c.h.m.i.r.r.CompositeRecordFileReader Loading record format version 5 from record file: 2021-01-14T20_03_18.004362000Z.rcd +2021-03-01T09:23:58.407-0600 INFO main c.h.m.i.p.r.RecordFileParserPerformanceTest Executing: parse() +2021-03-01T09:23:58.556-0600 INFO main c.h.m.i.c.MirrorDateRangePropertiesProcessor RECORD: parser will parse items in the range [1970-01-01T00:00:00Z, 2262-04-11T23:47:16.854775807Z] +2021-03-01T09:24:00.892-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 12654 rows to transaction table in 343.9 ms +2021-03-01T09:24:01.439-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 50588 rows to crypto_transfer table in 483.9 ms +2021-03-01T09:24:01.516-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 17 entities in 76.62 ms +2021-03-01T09:24:01.516-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 2.097 s +2021-03-01T09:24:01.573-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 12654 transactions from record file 2021-01-14T20_02_56.025429000Z.rcd in 3.118 s. Success: true +2021-03-01T09:24:02.475-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 12378 rows to transaction table in 546.3 ms +2021-03-01T09:24:03.535-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 49477 rows to crypto_transfer table in 1.039 s +2021-03-01T09:24:03.548-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 17 entities in 12.97 ms +2021-03-01T09:24:03.548-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.786 s +2021-03-01T09:24:03.550-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 12378 transactions from record file 2021-01-14T20_02_58.019176000Z.rcd in 1.892 s. Success: true +2021-03-01T09:24:04.588-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11649 rows to transaction table in 488.3 ms +2021-03-01T09:24:04.757-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 46575 rows to crypto_transfer table in 149.5 ms +2021-03-01T09:24:04.770-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 17 entities in 12.78 ms +2021-03-01T09:24:04.770-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.117 s +2021-03-01T09:24:04.771-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11649 transactions from record file 2021-01-14T20_03_00.032295000Z.rcd in 1.184 s. Success: true +2021-03-01T09:24:06.210-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11196 rows to transaction table in 1.325 s +2021-03-01T09:24:06.425-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 44753 rows to crypto_transfer table in 190.2 ms +2021-03-01T09:24:06.437-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 11.89 ms +2021-03-01T09:24:06.437-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.592 s +2021-03-01T09:24:06.439-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11196 transactions from record file 2021-01-14T20_03_02.031984000Z.rcd in 1.650 s. Success: true +2021-03-01T09:24:07.169-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 10500 rows to transaction table in 626.7 ms +2021-03-01T09:24:07.906-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 41979 rows to crypto_transfer table in 719.5 ms +2021-03-01T09:24:07.919-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 12.65 ms +2021-03-01T09:24:07.919-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.422 s +2021-03-01T09:24:07.921-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 10500 transactions from record file 2021-01-14T20_03_04.004057000Z.rcd in 1.473 s. Success: true +2021-03-01T09:24:08.607-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11427 rows to transaction table in 551.0 ms +2021-03-01T09:24:09.021-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 45665 rows to crypto_transfer table in 396.4 ms +2021-03-01T09:24:09.033-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 17 entities in 12.04 ms +2021-03-01T09:24:09.034-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 1.047 s +2021-03-01T09:24:09.035-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11427 transactions from record file 2021-01-14T20_03_06.001417000Z.rcd in 1.107 s. Success: true +2021-03-01T09:24:09.424-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11771 rows to transaction table in 115.9 ms +2021-03-01T09:24:09.804-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 47053 rows to crypto_transfer table in 358.5 ms +2021-03-01T09:24:09.816-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 11.80 ms +2021-03-01T09:24:09.816-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 634.3 ms +2021-03-01T09:24:09.817-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11771 transactions from record file 2021-01-14T20_03_08.002071000Z.rcd in 719.5 ms. Success: true +2021-03-01T09:24:10.178-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 10895 rows to transaction table in 230.7 ms +2021-03-01T09:24:10.523-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 43554 rows to crypto_transfer table in 327.1 ms +2021-03-01T09:24:10.535-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 11.07 ms +2021-03-01T09:24:10.535-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 645.9 ms +2021-03-01T09:24:10.537-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 10895 transactions from record file 2021-01-14T20_03_10.016188000Z.rcd in 711.1 ms. Success: true +2021-03-01T09:24:10.739-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 10642 rows to transaction table in 98.91 ms +2021-03-01T09:24:10.930-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 42537 rows to crypto_transfer table in 170.9 ms +2021-03-01T09:24:10.943-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 13.00 ms +2021-03-01T09:24:10.944-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 341.4 ms +2021-03-01T09:24:10.947-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 10642 transactions from record file 2021-01-14T20_03_12.001771000Z.rcd in 400.9 ms. Success: true +2021-03-01T09:24:13.375-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11801 rows to transaction table in 2.272 s +2021-03-01T09:24:13.747-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 47174 rows to crypto_transfer table in 353.5 ms +2021-03-01T09:24:13.761-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 12.79 ms +2021-03-01T09:24:13.761-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 2.722 s +2021-03-01T09:24:13.762-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11801 transactions from record file 2021-01-14T20_03_14.015461000Z.rcd in 2.808 s. Success: true +2021-03-01T09:24:14.249-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 11044 rows to transaction table in 302.3 ms +2021-03-01T09:24:14.583-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 44158 rows to crypto_transfer table in 321.3 ms +2021-03-01T09:24:14.602-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 10.98 ms +2021-03-01T09:24:14.602-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 694.9 ms +2021-03-01T09:24:14.603-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 11044 transactions from record file 2021-01-14T20_03_16.061108000Z.rcd in 801.3 ms. Success: true +2021-03-01T09:24:14.910-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 10665 rows to transaction table in 226.0 ms +2021-03-01T09:24:15.265-0600 INFO main c.h.m.i.p.r.e.s.PgCopy Copied 42630 rows to crypto_transfer table in 338.9 ms +2021-03-01T09:24:15.276-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Inserted 16 entities in 10.03 ms +2021-03-01T09:24:15.276-0600 INFO main c.h.m.i.p.r.e.s.SqlEntityListener Completed batch inserts in 618.4 ms +2021-03-01T09:24:15.277-0600 INFO main c.h.m.i.p.r.RecordFileParser Finished parsing 10665 transactions from record file 2021-01-14T20_03_18.004362000Z.rcd in 663.4 ms. Success: true +[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 49.831 s <<< FAILURE! - in com.hedera.mirror.importer.parser.record.RecordFileParserPerformanceTest +[ERROR] com.hedera.mirror.importer.parser.record.RecordFileParserPerformanceTest.parse Time elapsed: 17.606 s <<< ERROR! +java.util.concurrent.TimeoutException: parse() timed out after 15 seconds +``` + +**Actual Behavior** +Steps to reproduce the behavior: +1. Go to '...' +2. Click on '....' +3. Scroll down to '....' + +**Expected Behavior** + +**Environment:** +- Java: [e.g. OpenJDK 11.0.4] +- Node: [e.g. v12.9.1-x86] +- OS: [e.g. Ubuntu 18.04] +- Version: [e.g. 1.0.1] + +**Additional Context** + +",0,recordfileparserperformancetest times out in circleci thanks for submitting a bug report before submitting try searching the existing issues to see if your issue has already been reported try asking on our discord group if you re reporting a security vulnerability please email security hedera com instead of opening an issue detailed description info main c h m i p r recordfileparserperformancetest no active profile set falling back to default profiles default info main c h m i c metricsconfiguration collecting table metrics info main c h m i c credentialsproviderconfiguration setting up async client using aws default credentials provider info main o f c i c dbmigrate current version of schema public info main o f c i c dbmigrate migrating schema public to version init info main o f c i c dbmigrate migrating schema public to version update info main o f c i c dbmigrate migrating schema public to version procedures info main o f c i c dbmigrate migrating schema public to version procedures info main o f c i c dbmigrate migrating schema public to version update info main o f c i c dbmigrate migrating schema public to version file create patch info main o f c i c dbmigrate migrating schema public to version application status info main o f c i c dbmigrate migrating schema public to version index info main o f c i c dbmigrate migrating schema public to version transactions indexes info main o f c i c dbmigrate migrating schema public to version account balances info main o f c i c dbmigrate migrating schema public to version freeze transaction info main o f c i c dbmigrate migrating schema public to version response codes info main o f c i c dbmigrate migrating schema public to version account balances info main o f c i c dbmigrate migrating schema public to version entity key info main o f c i c dbmigrate migrating schema public to version remove deprecated info main o f c i c dbmigrate migrating schema public to version transactions pk info main o f c i c dbmigrate migrating schema public to version entity key info main o f c i c dbmigrate migrating schema public to version duplicate entities info main o f c i c dbmigrate migrating schema public to version entity create info main o f c i c dbmigrate migrating schema public to version response codes info main o f c i c dbmigrate migrating schema public to version missing entities warn main c h m i m missing entities skipping entity import due to missing file root project hedera mirror importer data accountinfo txt info main o f c i c dbmigrate migrating schema public to version transactions maxfee duration info main o f c i c dbmigrate migrating schema public to version transaction hash info main o f c i c dbmigrate migrating schema public to version response codes info main o f c i c dbmigrate migrating schema public to version unknown transaction types info main o f c i c dbmigrate migrating schema public to version cryptotransferlist realm num info main o f c i c dbmigrate migrating schema public to version hcs support info main o f c i c dbmigrate migrating schema public to version f entity create info main o f c i c dbmigrate migrating schema public to version transaction bytes info main o f c i c dbmigrate migrating schema public to version drop f entity create info main o f c i c dbmigrate migrating schema public to version hcs proto changes info main o f c i c dbmigrate migrating schema public to version fix mirror api permissions info main o f c i c dbmigrate migrating schema public to version non fee transfers info main o f c i c dbmigrate migrating schema public to version drop fks trigger unused indexes info main o f c i c dbmigrate migrating schema public to version drop file id info main o f c i c dbmigrate migrating schema public to version add topicmessage seqindex info main o f c i c dbmigrate migrating schema public to version remove unused components info main o f c i c dbmigrate migrating schema public to version proto updates info main o f c i c dbmigrate migrating schema public to version add topicmessage runninghashversion info main o f c i c dbmigrate migrating schema public to version remove bypass state info main o f c i c dbmigrate migrating schema public to version t record files timestamps info main o f c i c dbmigrate migrating schema public to version denormalize entities info main o f c i c dbmigrate migrating schema public to version denormalize transfer tables info main o f c i c dbmigrate migrating schema public to version topic message fragmentation info main o f c i c dbmigrate migrating schema public to version pg notify info main o f c i c dbmigrate migrating schema public to version drop notify trigger info main o f c i c dbmigrate migrating schema public to version add event streams info main o f c i c dbmigrate migrating schema public to version populate consensus start end for record files info main o f c i c dbmigrate migrating schema public to version update name of record file to filename info main o f c i c dbmigrate migrating schema public to version rename tables info main o f c i c dbmigrate migrating schema public to version address book info main o f c i c dbmigrate migrating schema public to version record file index consensus end info main o f c i c dbmigrate migrating schema public to version store verification node in stream file tables info main o f c i c dbmigrate migrating schema public to version balance entity id info main o f c i c dbmigrate migrating schema public to version hts support info main o f c i c dbmigrate migrating schema public to version fix address book info main o f c i c dbmigrate migrating schema public to version token symbol size info main o f c i c dbmigrate migrating schema public to version remove invalid entities info main o f c i c dbmigrate migrating schema public to version delete token transfers transaction type info main o f c i c dbmigrate migrating schema public to version update transaction index info main o f c i c dbmigrate migrating schema public to version drop token account id info main o f c i c dbmigrate migrating schema public to version update record file info main o f c i c dbmigrate migrating schema public to version schedule support info main o f c i c dbmigrate migrating schema public to version transaction schedule info main o f c i c dbmigrate migrating schema public to version rename record file hash info main o f c i c dbmigrate migrating schema public to version update schedule response codes info main o f c i c dbmigrate migrating schema public to version remove filesystem info main o f c i c dbmigrate migrating schema public to version schedule index info main o f c i c dbmigrate migrating schema public to version record file block index info main o f c i c dbmigrate successfully applied migrations to schema public execution time info main c h m i c mirrorimporterconfiguration configured to download from in region us east with bucket name test info main c h m i p r recordfileparserperformancetest started recordfileparserperformancetest in seconds jvm running for info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i r r compositerecordfilereader loading record format version from record file rcd info main c h m i p r recordfileparserperformancetest executing parse info main c h m i c mirrordaterangepropertiesprocessor record parser will parse items in the range info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in s info main c h m i p r recordfileparser finished parsing transactions from record file rcd in s success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in s info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in s info main c h m i p r recordfileparser finished parsing transactions from record file rcd in s success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in s info main c h m i p r recordfileparser finished parsing transactions from record file rcd in s success true info main c h m i p r e s pgcopy copied rows to transaction table in s info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in s info main c h m i p r recordfileparser finished parsing transactions from record file rcd in s success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in s info main c h m i p r recordfileparser finished parsing transactions from record file rcd in s success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in s info main c h m i p r recordfileparser finished parsing transactions from record file rcd in s success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in ms info main c h m i p r recordfileparser finished parsing transactions from record file rcd in ms success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in ms info main c h m i p r recordfileparser finished parsing transactions from record file rcd in ms success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in ms info main c h m i p r recordfileparser finished parsing transactions from record file rcd in ms success true info main c h m i p r e s pgcopy copied rows to transaction table in s info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in s info main c h m i p r recordfileparser finished parsing transactions from record file rcd in s success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in ms info main c h m i p r recordfileparser finished parsing transactions from record file rcd in ms success true info main c h m i p r e s pgcopy copied rows to transaction table in ms info main c h m i p r e s pgcopy copied rows to crypto transfer table in ms info main c h m i p r e s sqlentitylistener inserted entities in ms info main c h m i p r e s sqlentitylistener completed batch inserts in ms info main c h m i p r recordfileparser finished parsing transactions from record file rcd in ms success true tests run failures errors skipped time elapsed s failure in com hedera mirror importer parser record recordfileparserperformancetest com hedera mirror importer parser record recordfileparserperformancetest parse time elapsed s error java util concurrent timeoutexception parse timed out after seconds actual behavior steps to reproduce the behavior go to click on scroll down to expected behavior environment java node os version additional context ,0 +3380,6346700935.0,IssuesEvent,2017-07-28 03:27:37,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,opened,Built-in function SOUNDEX not supported,compatibility rc3.1,"Please answer these questions before submitting your issue. Thanks! + +1. What did you do? + +``` +create table t1 (id integer, a varchar(100) character set utf8 collate utf8_unicode_ci); +insert into t1 values (1, 'Test'); +select * from t1 where soundex(a) = soundex('Test'); +``` + +2. What did you expect to see? + +``` ++------+------+ +| id | a | ++------+------+ +| 1 | Test | ++------+------+ +1 row in set (0.01 sec) +``` + +3. What did you see instead? + +``` +ERROR 1105 (HY000): line 0 column 31 near ""a) = soundex('Test')"" (total length 51) +``` + +4. What version of TiDB are you using (`tidb-server -V`)? + +",True,"Built-in function SOUNDEX not supported - Please answer these questions before submitting your issue. Thanks! + +1. What did you do? + +``` +create table t1 (id integer, a varchar(100) character set utf8 collate utf8_unicode_ci); +insert into t1 values (1, 'Test'); +select * from t1 where soundex(a) = soundex('Test'); +``` + +2. What did you expect to see? + +``` ++------+------+ +| id | a | ++------+------+ +| 1 | Test | ++------+------+ +1 row in set (0.01 sec) +``` + +3. What did you see instead? + +``` +ERROR 1105 (HY000): line 0 column 31 near ""a) = soundex('Test')"" (total length 51) +``` + +4. What version of TiDB are you using (`tidb-server -V`)? + +",1,built in function soundex not supported please answer these questions before submitting your issue thanks what did you do create table id integer a varchar character set collate unicode ci insert into values test select from where soundex a soundex test what did you expect to see id a test row in set sec what did you see instead error line column near a soundex test total length what version of tidb are you using tidb server v ,1 +385592,26644872795.0,IssuesEvent,2023-01-25 09:11:43,ebi-ait/hca-ebi-wrangler-central,https://api.github.com/repos/ebi-ait/hca-ebi-wrangler-central,closed,SOP - Inform the contributors after wrangling,documentation operations,"To increase community engagement, we should send an email after wrangling any dataset from a publication or repository (GEO). After verifying the wrangled dataset is live on the Data Portal, we should inform about it the contributors. +Add a new template email: after verifying a dataset is live on the Data Portal, we should inform the original contributors of the dataset. A good practice would be to ask them for feedback and to join the conversation on slack. -Dependency Hierarchy: - - dotnetcore-sample-1.0.0 (Root Library) - - microsoft.aspnetcore.2.0.0.nupkg - - :x: **microsoft.aspnetcore.server.iisintegration.2.0.0.nupkg** (Vulnerable Library) +This idea comes from one of our DCP meetings, to increase community engagement and as a wrangler's best practice +Completed when: -### microsoft.aspnetcore.hosting.2.0.0.nupkg
+- [x] A new template email has been created +- [x] Our SOPs are updated to reflect this process",1.0,"SOP - Inform the contributors after wrangling - To increase community engagement, we should send an email after wrangling any dataset from a publication or repository (GEO). After verifying the wrangled dataset is live on the Data Portal, we should inform about it the contributors. +Add a new template email: after verifying a dataset is live on the Data Portal, we should inform the original contributors of the dataset. A good practice would be to ask them for feedback and to join the conversation on slack. -ASP.NET Core hosting infrastructure and startup logic for web applications.
-Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.hosting.2.0.0.nupkg
-Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
-Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.hosting/2.0.0/microsoft.aspnetcore.hosting.2.0.0.nupkg
-+This idea comes from one of our DCP meetings, to increase community engagement and as a wrangler's best practice -Dependency Hierarchy: - - dotnetcore-sample-1.0.0 (Root Library) - - microsoft.aspnetcore.2.0.0.nupkg - - microsoft.aspnetcore.server.kestrel.2.0.0.nupkg - - :x: **microsoft.aspnetcore.hosting.2.0.0.nupkg** (Vulnerable Library) +Completed when: +- [x] A new template email has been created +- [x] Our SOPs are updated to reflect this process",0,sop inform the contributors after wrangling to increase community engagement we should send an email after wrangling any dataset from a publication or repository geo after verifying the wrangled dataset is live on the data portal we should inform about it the contributors add a new template email after verifying a dataset is live on the data portal we should inform the original contributors of the dataset a good practice would be to ask them for feedback and to join the conversation on slack this idea comes from one of our dcp meetings to increase community engagement and as a wrangler s best practice completed when a new template email has been created our sops are updated to reflect this process,0 +11634,5052548446.0,IssuesEvent,2016-12-21 02:48:53,rust-lang/rust,https://api.github.com/repos/rust-lang/rust,closed,rustbuild: Dist on cross builds produce way too many tarballs,A-rustbuild,"The [dist step](https://buildbot.rust-lang.org/builders/nightly-dist-rustc-cross-host-linux/builds/281/steps/distcheck/logs/stdio) on cross builds takes over half an hour, and that's a crazy amount of time for something which should take at most 5 minutes. + +The [list of tarballs created](https://gist.github.com/alexcrichton/0ef4bdfcb3514e1ff5362394056feb4d) is massively duplicated. Namely we create around 60 source tarballs (!!). Looks like we're also packaging up the standard library dozens of times for the same target. + +My guess as to what's happening is that by default rustbuild is doing a cross product of (host x target) and doing all the builds. We should skip a massive number of them for distribution and ideally not even put them into the step list at all! + +Note that this isn't a fire at the moment as everything's just overwriting what was previously created, but this sounds like something that *could* become a fire! + +Should also be a relatively easy bug to knock out for anyone who wants to become familiar with rustbuild!",1.0,"rustbuild: Dist on cross builds produce way too many tarballs - The [dist step](https://buildbot.rust-lang.org/builders/nightly-dist-rustc-cross-host-linux/builds/281/steps/distcheck/logs/stdio) on cross builds takes over half an hour, and that's a crazy amount of time for something which should take at most 5 minutes. + +The [list of tarballs created](https://gist.github.com/alexcrichton/0ef4bdfcb3514e1ff5362394056feb4d) is massively duplicated. Namely we create around 60 source tarballs (!!). Looks like we're also packaging up the standard library dozens of times for the same target. + +My guess as to what's happening is that by default rustbuild is doing a cross product of (host x target) and doing all the builds. We should skip a massive number of them for distribution and ideally not even put them into the step list at all! + +Note that this isn't a fire at the moment as everything's just overwriting what was previously created, but this sounds like something that *could* become a fire! + +Should also be a relatively easy bug to knock out for anyone who wants to become familiar with rustbuild!",0,rustbuild dist on cross builds produce way too many tarballs the on cross builds takes over half an hour and that s a crazy amount of time for something which should take at most minutes the is massively duplicated namely we create around source tarballs looks like we re also packaging up the standard library dozens of times for the same target my guess as to what s happening is that by default rustbuild is doing a cross product of host x target and doing all the builds we should skip a massive number of them for distribution and ideally not even put them into the step list at all note that this isn t a fire at the moment as everything s just overwriting what was previously created but this sounds like something that could become a fire should also be a relatively easy bug to knock out for anyone who wants to become familiar with rustbuild ,0 +11105,13130760190.0,IssuesEvent,2020-08-06 15:51:17,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,support string convert to number in best effort during compare.,status/help-wanted type/compatibility,"1. What did you do? +``` +CREATE TABLE `test3` (`id` int(11) DEFAULT NULL); +CREATE TABLE `test4` (`k` varchar(3) DEFAULT NULL); + +insert into test3 values (1); +insert into test4 values (""1-2""); + +select * from test3 a inner join test4 b on a.id = b.k; +``` + + +2. What did you expect to see? + +``` +mysql> select * from test3 a inner join test4 b on a.id = b.k; ++------+------+ +| id | k | ++------+------+ +| 1 | 1-2 | ++------+------+ +1 row in set (0.00 sec) +``` + +3. What did you see instead? + +``` +ERROR 1105 (HY000): strconv.ParseFloat: parsing ""1-2"": invalid syntax +``` + +4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? + +``` +| Release Version: v2.0.0 +Git Commit Hash: 637e130e6a9ba2e54e158131c0466233db39a60e +Git Branch: release-2.0 +UTC Build Time: 2018-04-27 11:43:00 +GoVersion: go version go1.10 linux/amd64 +TiKV Min Version: 2.0.0-rc.4.1 | +``` +",True,"support string convert to number in best effort during compare. - 1. What did you do? +``` +CREATE TABLE `test3` (`id` int(11) DEFAULT NULL); +CREATE TABLE `test4` (`k` varchar(3) DEFAULT NULL); + +insert into test3 values (1); +insert into test4 values (""1-2""); + +select * from test3 a inner join test4 b on a.id = b.k; +``` + + +2. What did you expect to see? + +``` +mysql> select * from test3 a inner join test4 b on a.id = b.k; ++------+------+ +| id | k | ++------+------+ +| 1 | 1-2 | ++------+------+ +1 row in set (0.00 sec) +``` + +3. What did you see instead? + +``` +ERROR 1105 (HY000): strconv.ParseFloat: parsing ""1-2"": invalid syntax +``` + +4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? + +``` +| Release Version: v2.0.0 +Git Commit Hash: 637e130e6a9ba2e54e158131c0466233db39a60e +Git Branch: release-2.0 +UTC Build Time: 2018-04-27 11:43:00 +GoVersion: go version go1.10 linux/amd64 +TiKV Min Version: 2.0.0-rc.4.1 | +``` +",1,support string convert to number in best effort during compare what did you do create table id int default null create table k varchar default null insert into values insert into values select from a inner join b on a id b k what did you expect to see mysql select from a inner join b on a id b k id k row in set sec what did you see instead error strconv parsefloat parsing invalid syntax what version of tidb are you using tidb server v or run select tidb version on tidb release version git commit hash git branch release utc build time goversion go version linux tikv min version rc ,1 +16233,21796759895.0,IssuesEvent,2022-05-15 18:53:51,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,TiDB gives a different result with `SET GLOBAL table_open_cache=-1;` ,type/compatibility,"## Bug Report + + +### 1. Minimal reproduce step +```sql +SET GLOBAL table_open_cache=-1; +SHOW VARIABLES LIKE 'table_open_cache'; +``` + +### 2. What did you expect to see? +```sql +mysql> SHOW VARIABLES LIKE 'table_open_cache'; ++------------------+-------+ +| Variable_name | Value | ++------------------+-------+ +| table_open_cache | 1 | ++------------------+-------+ +1 row in set (0.03 sec) +``` + +### 3. What did you see instead +```sql +TiDB> SHOW VARIABLES LIKE 'table_open_cache'; ++------------------+-------+ +| Variable_name | Value | ++------------------+-------+ +| table_open_cache | -1 | ++------------------+-------+ +1 row in set (0.01 sec) +``` + +### 4. What is your TiDB version? +```sql +tidb_version(): Release Version: v5.2.2 +Edition: Community +Git Commit Hash: da1c21fd45a4ea5900ac16d2f4a248143f378d18 +Git Branch: heads/refs/tags/v5.2.2 +UTC Build Time: 2021-10-20 06:03:29 +GoVersion: go1.16.4 +Race Enabled: false +TiKV Min Version: v3.0.0-60965b006877ca7234adaced7890d7b029ed1306 +Check Table Before Drop: false +``` + +",True,"TiDB gives a different result with `SET GLOBAL table_open_cache=-1;` - ## Bug Report + + +### 1. Minimal reproduce step +```sql +SET GLOBAL table_open_cache=-1; +SHOW VARIABLES LIKE 'table_open_cache'; +``` + +### 2. What did you expect to see? +```sql +mysql> SHOW VARIABLES LIKE 'table_open_cache'; ++------------------+-------+ +| Variable_name | Value | ++------------------+-------+ +| table_open_cache | 1 | ++------------------+-------+ +1 row in set (0.03 sec) +``` + +### 3. What did you see instead +```sql +TiDB> SHOW VARIABLES LIKE 'table_open_cache'; ++------------------+-------+ +| Variable_name | Value | ++------------------+-------+ +| table_open_cache | -1 | ++------------------+-------+ +1 row in set (0.01 sec) +``` + +### 4. What is your TiDB version? +```sql +tidb_version(): Release Version: v5.2.2 +Edition: Community +Git Commit Hash: da1c21fd45a4ea5900ac16d2f4a248143f378d18 +Git Branch: heads/refs/tags/v5.2.2 +UTC Build Time: 2021-10-20 06:03:29 +GoVersion: go1.16.4 +Race Enabled: false +TiKV Min Version: v3.0.0-60965b006877ca7234adaced7890d7b029ed1306 +Check Table Before Drop: false +``` + +",1,tidb gives a different result with set global table open cache bug report minimal reproduce step sql set global table open cache show variables like table open cache what did you expect to see sql mysql show variables like table open cache variable name value table open cache row in set sec what did you see instead sql tidb show variables like table open cache variable name value table open cache row in set sec what is your tidb version sql tidb version release version edition community git commit hash git branch heads refs tags utc build time goversion race enabled false tikv min version check table before drop false ,1 +11070,13097298340.0,IssuesEvent,2020-08-03 17:12:02,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,The result of `show variables` is not compatible with MySQL,status/help-wanted type/compatibility type/duplicate,"## Bug Report + +Please answer these questions before submitting your issue. Thanks! + +1. What did you do? +If possible, provide a recipe for reproducing the error. +``` +select @@global.sql_log_bin; +show variables like ""sql_log_bin""; +``` + +2. What did you expect to see? +Your MySQL connection id is 172 +Server version: 5.7.20 Homebrew +``` +mysql> select @@global.sql_log_bin; ++----------------------+ +| @@global.sql_log_bin | ++----------------------+ +| 1 | ++----------------------+ +1 row in set, 1 warning (0.00 sec) + +mysql> show variables like ""sql_log_bin""; ++---------------+-------+ +| Variable_name | Value | ++---------------+-------+ +| sql_log_bin | ON | ++---------------+-------+ +1 row in set (0.00 sec) +``` + + +3. What did you see instead? +``` +tidb> select @@global.sql_log_bin; ++----------------------+ +| @@global.sql_log_bin | ++----------------------+ +| 1 | ++----------------------+ +1 row in set (0.00 sec) + +tidb> show variables like ""sql_log_bin""; ++---------------+-------+ +| Variable_name | Value | ++---------------+-------+ +| sql_log_bin | 1 | ++---------------+-------+ +1 row in set (0.02 sec) +``` + + +4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? +Release Version: v3.0.0-beta-69-g84d1513b0-dirty +Git Commit Hash: 84d1513b09222c60ffba2d3397b6745e5148a603 +Git Branch: master",True,"The result of `show variables` is not compatible with MySQL - ## Bug Report + +Please answer these questions before submitting your issue. Thanks! + +1. What did you do? +If possible, provide a recipe for reproducing the error. +``` +select @@global.sql_log_bin; +show variables like ""sql_log_bin""; +``` + +2. What did you expect to see? +Your MySQL connection id is 172 +Server version: 5.7.20 Homebrew +``` +mysql> select @@global.sql_log_bin; ++----------------------+ +| @@global.sql_log_bin | ++----------------------+ +| 1 | ++----------------------+ +1 row in set, 1 warning (0.00 sec) + +mysql> show variables like ""sql_log_bin""; ++---------------+-------+ +| Variable_name | Value | ++---------------+-------+ +| sql_log_bin | ON | ++---------------+-------+ +1 row in set (0.00 sec) +``` + + +3. What did you see instead? +``` +tidb> select @@global.sql_log_bin; ++----------------------+ +| @@global.sql_log_bin | ++----------------------+ +| 1 | ++----------------------+ +1 row in set (0.00 sec) + +tidb> show variables like ""sql_log_bin""; ++---------------+-------+ +| Variable_name | Value | ++---------------+-------+ +| sql_log_bin | 1 | ++---------------+-------+ +1 row in set (0.02 sec) +``` + + +4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? +Release Version: v3.0.0-beta-69-g84d1513b0-dirty +Git Commit Hash: 84d1513b09222c60ffba2d3397b6745e5148a603 +Git Branch: master",1,the result of show variables is not compatible with mysql bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error select global sql log bin show variables like sql log bin what did you expect to see your mysql connection id is server version homebrew mysql select global sql log bin global sql log bin row in set warning sec mysql show variables like sql log bin variable name value sql log bin on row in set sec what did you see instead tidb select global sql log bin global sql log bin row in set sec tidb show variables like sql log bin variable name value sql log bin row in set sec what version of tidb are you using tidb server v or run select tidb version on tidb release version beta dirty git commit hash git branch master,1 +136366,11047445288.0,IssuesEvent,2019-12-09 18:59:23,dexpenses/dexpenses-extract,https://api.github.com/repos/dexpenses/dexpenses-extract,closed,Implement test receipt normal/dransfeld-rewe-debit,enhancement test-data,"Receipt to implement: +",1.0,"Implement test receipt normal/dransfeld-rewe-debit - Receipt to implement: +",0,implement test receipt normal dransfeld rewe debit receipt to implement normal dransfeld rewe debit ,0 +7181,9473698991.0,IssuesEvent,2019-04-19 03:23:17,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,not support show open table ,help wanted type/compatibility,"## Bug Report + +1. What did you do? +If possible, provide a recipe for reproducing the error. + +I use heidisql, and meet error caused by unsupported `show open tables` command. + + +2. What did you expect to see? + +no error + +3. What did you see instead? + +``` +ERROR 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your TiDB version for the right syntax to use line 1 column 9 near ""OPEN TABLES"" +``` + +4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? + +",True,"not support show open table - ## Bug Report + +1. What did you do? +If possible, provide a recipe for reproducing the error. + +I use heidisql, and meet error caused by unsupported `show open tables` command. + + +2. What did you expect to see? + +no error + +3. What did you see instead? + +``` +ERROR 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your TiDB version for the right syntax to use line 1 column 9 near ""OPEN TABLES"" +``` + +4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? + +",1,not support show open table bug report what did you do if possible provide a recipe for reproducing the error i use heidisql and meet error caused by unsupported show open tables command what did you expect to see no error what did you see instead error you have an error in your sql syntax check the manual that corresponds to your tidb version for the right syntax to use line column near open tables what version of tidb are you using tidb server v or run select tidb version on tidb ,1 +3637,6524200471.0,IssuesEvent,2017-08-29 11:45:09,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,Incorrect Decimal inference for float literal,bug compatibility for-new-contributors help wanted,"MySQL: + +``` +mysql> SELECT 151113102019e0; +Field 1: `151113102019e0` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: DOUBLE +Collation: binary (63) +Length: 14 +Max_length: 12 +Decimals: 31 +Flags: NOT_NULL BINARY NUM + + ++----------------+ +| 151113102019e0 | ++----------------+ +| 151113102019 | ++----------------+ +1 row in set (0.00 sec) + +mysql> SELECT 15111310201912e-2; +Field 1: `15111310201912e-2` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: DOUBLE +Collation: binary (63) +Length: 17 +Max_length: 15 +Decimals: 31 +Flags: NOT_NULL BINARY NUM + + ++-------------------+ +| 15111310201912e-2 | ++-------------------+ +| 151113102019.12 | ++-------------------+ +1 row in set (0.00 sec) +``` + +TiDB: + +``` +mysql> SELECT 151113102019e0; +Field 1: `151113102019e0` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: DOUBLE +Collation: binary (63) +Length: 12 +Max_length: 25 +Decimals: 12 +Flags: BINARY NUM + + ++---------------------------+ +| 151113102019e0 | ++---------------------------+ +| 151113102019.000000000000 | ++---------------------------+ +1 row in set (0.00 sec) + +mysql> SELECT 15111310201912e-2; +Field 1: `15111310201912e-2` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: DOUBLE +Collation: binary (63) +Length: 15 +Max_length: 15 +Decimals: 2 +Flags: BINARY NUM + + ++-------------------+ +| 15111310201912e-2 | ++-------------------+ +| 151113102019.12 | ++-------------------+ +1 row in set (0.00 sec) +```",True,"Incorrect Decimal inference for float literal - MySQL: + +``` +mysql> SELECT 151113102019e0; +Field 1: `151113102019e0` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: DOUBLE +Collation: binary (63) +Length: 14 +Max_length: 12 +Decimals: 31 +Flags: NOT_NULL BINARY NUM + + ++----------------+ +| 151113102019e0 | ++----------------+ +| 151113102019 | ++----------------+ +1 row in set (0.00 sec) + +mysql> SELECT 15111310201912e-2; +Field 1: `15111310201912e-2` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: DOUBLE +Collation: binary (63) +Length: 17 +Max_length: 15 +Decimals: 31 +Flags: NOT_NULL BINARY NUM + + ++-------------------+ +| 15111310201912e-2 | ++-------------------+ +| 151113102019.12 | ++-------------------+ +1 row in set (0.00 sec) +``` + +TiDB: + +``` +mysql> SELECT 151113102019e0; +Field 1: `151113102019e0` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: DOUBLE +Collation: binary (63) +Length: 12 +Max_length: 25 +Decimals: 12 +Flags: BINARY NUM + + ++---------------------------+ +| 151113102019e0 | ++---------------------------+ +| 151113102019.000000000000 | ++---------------------------+ +1 row in set (0.00 sec) + +mysql> SELECT 15111310201912e-2; +Field 1: `15111310201912e-2` +Catalog: `def` +Database: `` +Table: `` +Org_table: `` +Type: DOUBLE +Collation: binary (63) +Length: 15 +Max_length: 15 +Decimals: 2 +Flags: BINARY NUM + + ++-------------------+ +| 15111310201912e-2 | ++-------------------+ +| 151113102019.12 | ++-------------------+ +1 row in set (0.00 sec) +```",1,incorrect decimal inference for float literal mysql mysql select field catalog def database table org table type double collation binary length max length decimals flags not null binary num row in set sec mysql select field catalog def database table org table type double collation binary length max length decimals flags not null binary num row in set sec tidb mysql select field catalog def database table org table type double collation binary length max length decimals flags binary num row in set sec mysql select field catalog def database table org table type double collation binary length max length decimals flags binary num row in set sec ,1 +7853,10062951627.0,IssuesEvent,2019-07-23 03:29:31,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,alter database syntax support.,type/compatibility,"TiDB: + +``` +mysql> alter database test1 character set utf8 COLLATE uft8_general_ci; +ERROR 1105 (HY000): line 0 column 14 near "" test1 character set utf8 COLLATE uft8_general_ci"" (total length 63) +mysql> alter database test1 character set utf8 COLLATE uft8_general_ci; +ERROR 1105 (HY000): line 0 column 14 near "" test1 character set utf8 COLLATE uft8_general_ci"" (total length 63) +``` + +see https://dev.mysql.com/doc/refman/5.7/en/alter-database.html",True,"alter database syntax support. - TiDB: + +``` +mysql> alter database test1 character set utf8 COLLATE uft8_general_ci; +ERROR 1105 (HY000): line 0 column 14 near "" test1 character set utf8 COLLATE uft8_general_ci"" (total length 63) +mysql> alter database test1 character set utf8 COLLATE uft8_general_ci; +ERROR 1105 (HY000): line 0 column 14 near "" test1 character set utf8 COLLATE uft8_general_ci"" (total length 63) +``` + +see https://dev.mysql.com/doc/refman/5.7/en/alter-database.html",1,alter database syntax support tidb mysql alter database character set collate general ci error line column near character set collate general ci total length mysql alter database character set collate general ci error line column near character set collate general ci total length see ,1 +18496,25582848919.0,IssuesEvent,2022-12-01 06:40:03,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,BR: 6.2.0 system table mysql.user can't be restored to 6.4.0,type/bug compatibility-breaker severity/critical component/br affects-6.2 affects-6.3 affects-6.4,"## Bug Report + +Please answer these questions before submitting your issue. Thanks! + +### 1. Minimal reproduce step (Required) +tidb changed mysql.user schema in 6.4.0 +version95 add a column `User_attributes` to `mysql.user` +version98 add a column `Token_issuer` to `mysql.user` + +when restore 6.2.0 mysql.user to 6.4.0 will report : +Error: column count mismatch, table: user, col in cluster: 38, col in backup: 37: [BR:Restore:ErrR + +### 2. What did you expect to see? (Required) +restore success +### 3. What did you see instead (Required) +Error: column count mismatch, table: user, col in cluster: 38, col in backup: 37: [BR:Restore:ErrR +### 4. What is your TiDB version? (Required) + +master + +",True,"BR: 6.2.0 system table mysql.user can't be restored to 6.4.0 - ## Bug Report + +Please answer these questions before submitting your issue. Thanks! + +### 1. Minimal reproduce step (Required) +tidb changed mysql.user schema in 6.4.0 +version95 add a column `User_attributes` to `mysql.user` +version98 add a column `Token_issuer` to `mysql.user` + +when restore 6.2.0 mysql.user to 6.4.0 will report : +Error: column count mismatch, table: user, col in cluster: 38, col in backup: 37: [BR:Restore:ErrR + +### 2. What did you expect to see? (Required) +restore success +### 3. What did you see instead (Required) +Error: column count mismatch, table: user, col in cluster: 38, col in backup: 37: [BR:Restore:ErrR +### 4. What is your TiDB version? (Required) + +master + +",1,br system table mysql user can t be restored to bug report please answer these questions before submitting your issue thanks minimal reproduce step required tidb changed mysql user schema in add a column user attributes to mysql user add a column token issuer to mysql user when restore mysql user to will report error column count mismatch table user col in cluster col in backup br restore errr what did you expect to see required restore success what did you see instead required error column count mismatch table user col in cluster col in backup br restore errr what is your tidb version required master ,1 +237379,18158501210.0,IssuesEvent,2021-09-27 06:44:44,aravindm711/ESNSE,https://api.github.com/repos/aravindm711/ESNSE,opened,Abstract for search engine,documentation,"Things to be done, +- Problem statement +- Goals +- Implementation +- Tools and technologies +",1.0,"Abstract for search engine - Things to be done, +- Problem statement +- Goals +- Implementation +- Tools and technologies +",0,abstract for search engine things to be done problem statement goals implementation tools and technologies ,0 +230317,7606824554.0,IssuesEvent,2018-04-30 14:36:10,openshiftio/openshift.io,https://api.github.com/repos/openshiftio/openshift.io,closed,Some user accounts are seeing a continuous stream of - f8osoproxy 401 errors,SEV2-high area/auth priority/P0 team/auth type/bug,"This problem started on my user account on April 27. After logging in, I was unable to create a new quickstart. + +The Chrome console showed a continuous stream of these errors: +``` +WebSocket connection to 'wss://f8osoproxy-test-dsaas- +production.09b5.dsaas.openshiftapps.com/oapi/v1/namespaces/ldimaggi/builds? +watch=true&access_token=...' failed: HTTP Authentication failed; no valid credentials available +``` +NOTE: In the above example - I removed the text of the token and username. + +The resolution for the problem was to refresh the OpenShift Online connection: + + +The cause for the problem is unclear - is this happening to older accounts because a token is expiring? My account was created on: 2017-04-24 + + + +",1.0,"Some user accounts are seeing a continuous stream of - f8osoproxy 401 errors - This problem started on my user account on April 27. After logging in, I was unable to create a new quickstart. + +The Chrome console showed a continuous stream of these errors: +``` +WebSocket connection to 'wss://f8osoproxy-test-dsaas- +production.09b5.dsaas.openshiftapps.com/oapi/v1/namespaces/ldimaggi/builds? +watch=true&access_token=...' failed: HTTP Authentication failed; no valid credentials available +``` +NOTE: In the above example - I removed the text of the token and username. + +The resolution for the problem was to refresh the OpenShift Online connection: + + +The cause for the problem is unclear - is this happening to older accounts because a token is expiring? My account was created on: 2017-04-24 + + + +",0,some user accounts are seeing a continuous stream of errors this problem started on my user account on april after logging in i was unable to create a new quickstart the chrome console showed a continuous stream of these errors websocket connection to wss test dsaas production dsaas openshiftapps com oapi namespaces ldimaggi builds watch true access token failed http authentication failed no valid credentials available note in the above example i removed the text of the token and username the resolution for the problem was to refresh the openshift online connection the cause for the problem is unclear is this happening to older accounts because a token is expiring my account was created on ,0 +86695,8047419660.0,IssuesEvent,2018-08-01 00:33:38,ipfs/go-ipfs,https://api.github.com/repos/ipfs/go-ipfs,closed,`foofoo.block` has more links than UnixFS `Blocksizes`,testing,"I'm refactoring the DAG reader (https://github.com/ipfs/go-ipfs/pull/5257) and I'm having an [error](https://github.com/ipfs/go-ipfs/pull/5257/files#diff-0046ca7d126fec814fde6c7ffc49ab0dR277) in the test file [`t0110-gateway.sh`](https://github.com/ipfs/go-ipfs/blob/e71dce5dfba3616e5ade4abf0844925389497a56/test/sharness/t0110-gateway.sh#L164) where the [`foofoo.block`](https://github.com/ipfs/go-ipfs/blob/e71dce5dfba3616e5ade4abf0844925389497a56/test/sharness/t0110-gateway.sh#L160) file, which is saved as a block and retrieved as a file DAG through the gateway, would appear to have an inconsistency between its number of DAG links and its number of UnixFS `Blocksizes`, is this expected in the test? If not, how could I regenerate that file with the UnixFS format fixed? + +/cc @magik6k ",1.0,"`foofoo.block` has more links than UnixFS `Blocksizes` - I'm refactoring the DAG reader (https://github.com/ipfs/go-ipfs/pull/5257) and I'm having an [error](https://github.com/ipfs/go-ipfs/pull/5257/files#diff-0046ca7d126fec814fde6c7ffc49ab0dR277) in the test file [`t0110-gateway.sh`](https://github.com/ipfs/go-ipfs/blob/e71dce5dfba3616e5ade4abf0844925389497a56/test/sharness/t0110-gateway.sh#L164) where the [`foofoo.block`](https://github.com/ipfs/go-ipfs/blob/e71dce5dfba3616e5ade4abf0844925389497a56/test/sharness/t0110-gateway.sh#L160) file, which is saved as a block and retrieved as a file DAG through the gateway, would appear to have an inconsistency between its number of DAG links and its number of UnixFS `Blocksizes`, is this expected in the test? If not, how could I regenerate that file with the UnixFS format fixed? + +/cc @magik6k ",0, foofoo block has more links than unixfs blocksizes i m refactoring the dag reader and i m having an in the test file where the file which is saved as a block and retrieved as a file dag through the gateway would appear to have an inconsistency between its number of dag links and its number of unixfs blocksizes is this expected in the test if not how could i regenerate that file with the unixfs format fixed cc ,0 +11080,13100718142.0,IssuesEvent,2020-08-04 01:27:46,pingcap/tidb,https://api.github.com/repos/pingcap/tidb,closed,The result of Function COALESCE is different from MySQL. ,component/expression status/help-wanted type/compatibility,"## Bug Report + +Please answer these questions before submitting your issue. Thanks! + +1. What did you do? +If possible, provide a recipe for reproducing the error. + +```mysql +create table test2 ( a decimal(40,20) unsigned); +insert into test2 value(12.98731000000000000000); +select cast(COALESCE(a, 'xxx' ) as unsigned) from test2; +``` + +2. What did you expect to see? + +In MySQL, +```mysql +mysql> select cast(COALESCE(a, 'xxx' ) as unsigned) from test2; ++---------------------------------------+ +| cast(COALESCE(a, 'xxx' ) as unsigned) | ++---------------------------------------+ +| 12 | ++---------------------------------------+ +1 row in set, 1 warning (0.00 sec) + +mysql> show warnings; ++---------+------+--------------------------------------------------------------+ +| Level | Code | Message | ++---------+------+--------------------------------------------------------------+ +| Warning | 1292 | Truncated incorrect INTEGER value: '12.98731000000000000000' | ++---------+------+--------------------------------------------------------------+ +1 row in set (0.00 sec) +``` + +3. What did you see instead? + +In TiDB, +```mysql +mysql> select cast(COALESCE(a, 'xxx' ) as unsigned) from test2; ++---------------------------------------+ +| cast(COALESCE(a, 'xxx' ) as unsigned) | ++---------------------------------------+ +| 13 | ++---------------------------------------+ +1 row in set (0.01 sec) +``` + + +4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? +8c20289c7",True,"The result of Function COALESCE is different from MySQL. - ## Bug Report + +Please answer these questions before submitting your issue. Thanks! + +1. What did you do? +If possible, provide a recipe for reproducing the error. + +```mysql +create table test2 ( a decimal(40,20) unsigned); +insert into test2 value(12.98731000000000000000); +select cast(COALESCE(a, 'xxx' ) as unsigned) from test2; +``` + +2. What did you expect to see? + +In MySQL, +```mysql +mysql> select cast(COALESCE(a, 'xxx' ) as unsigned) from test2; ++---------------------------------------+ +| cast(COALESCE(a, 'xxx' ) as unsigned) | ++---------------------------------------+ +| 12 | ++---------------------------------------+ +1 row in set, 1 warning (0.00 sec) + +mysql> show warnings; ++---------+------+--------------------------------------------------------------+ +| Level | Code | Message | ++---------+------+--------------------------------------------------------------+ +| Warning | 1292 | Truncated incorrect INTEGER value: '12.98731000000000000000' | ++---------+------+--------------------------------------------------------------+ +1 row in set (0.00 sec) +``` + +3. What did you see instead? + +In TiDB, +```mysql +mysql> select cast(COALESCE(a, 'xxx' ) as unsigned) from test2; ++---------------------------------------+ +| cast(COALESCE(a, 'xxx' ) as unsigned) | ++---------------------------------------+ +| 13 | ++---------------------------------------+ +1 row in set (0.01 sec) +``` + + +4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? +8c20289c7",1,the result of function coalesce is different from mysql bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error mysql create table a decimal unsigned insert into value select cast coalesce a xxx as unsigned from what did you expect to see in mysql mysql mysql select cast coalesce a xxx as unsigned from cast coalesce a xxx as unsigned row in set warning sec mysql show warnings level code message warning truncated incorrect integer value row in set sec what did you see instead in tidb mysql mysql select cast coalesce a xxx as unsigned from cast coalesce a xxx as unsigned row in set sec what version of tidb are you using tidb server v or run select tidb version on tidb ,1 +665237,22304417143.0,IssuesEvent,2022-06-13 11:45:18,ICST-Technion/WelliBe,https://api.github.com/repos/ICST-Technion/WelliBe,closed,Create a page where the user can send thank you notes to its doctors,medium priority feature,slides 16-20 in the presentation.,1.0,Create a page where the user can send thank you notes to its doctors - slides 16-20 in the presentation.,0,create a page where the user can send thank you notes to its doctors slides in the presentation ,0 +143939,22205566780.0,IssuesEvent,2022-06-07 14:37:38,patternfly/patternfly-org,https://api.github.com/repos/patternfly/patternfly-org,opened,Create guidelines for the Panel component,PF4 design Guidelines,"https://www.patternfly.org/v4/components/panel/design-guidelines -
Found in HEAD commit: d70903cc2e5e4b5d95bd49d679c3d9742e0c4fb9
-Found in base branch: main
- +Add guidelines including: +- what is a panel +- when to use a panel +- examples of panel in use - +Also mention that it works with the primary detail component, selectable tables etc....",1.0,"Create guidelines for the Panel component - https://www.patternfly.org/v4/components/panel/design-guidelines -### Vulnerability Details -- -ASP.NET Core 1.0. 1.1, and 2.0 allow an elevation of privilege vulnerability due to how ASP.NET web applications handle web requests, aka ""ASP.NET Core Elevation Of Privilege Vulnerability"". This CVE is unique from CVE-2018-0784. +Add guidelines including: +- what is a panel +- when to use a panel +- examples of panel in use -
Publish Date: 2018-03-14 -
URL: CVE-2018-0808
- +Also mention that it works with the primary detail component, selectable tables etc....",0,create guidelines for the panel component add guidelines including what is a panel when to use a panel examples of panel in use also mention that it works with the primary detail component selectable tables etc ,0 +184976,14291776599.0,IssuesEvent,2020-11-23 23:25:29,github-vet/rangeclosure-findings,https://api.github.com/repos/github-vet/rangeclosure-findings,closed,tuuuz/etcd: tests/integration/clientv3/txn_test.go; 23 LoC,fresh small test," +Found a possible issue in [tuuuz/etcd](https://www.github.com/tuuuz/etcd) at [tests/integration/clientv3/txn_test.go](https://github.com/tuuuz/etcd/blob/31a1226ec4edcd20180090f9a45250f29331bf89/tests/integration/clientv3/txn_test.go#L118-L140) - +The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements +which capture loop variables. -### CVSS 3 Score Details (7.5) -+[Click here to see the code in its original context.](https://github.com/tuuuz/etcd/blob/31a1226ec4edcd20180090f9a45250f29331bf89/tests/integration/clientv3/txn_test.go#L118-L140) -Base Score Metrics: -- Exploitability Metrics: - - Attack Vector: Network - - Attack Complexity: Low - - Privileges Required: None - - User Interaction: None - - Scope: Unchanged -- Impact Metrics: - - Confidentiality Impact: None - - Integrity Impact: None - - Availability Impact: High -
-For more information on CVSS3 Scores, click here. - ++ donec := make(chan struct{}, 1) + go func() { + _, err := kv.Txn(context.TODO()).Then(thenOps[i]...).Commit() + if err != nil { + t.Errorf(""expected response, got error %v"", err) + } + donec <- struct{}{} + }() + // wait for txn to fail on disconnect + time.Sleep(100 * time.Millisecond) -
Type: Upgrade version
-Origin: https://portal.msrc.microsoft.com/en-US/security-guidance/advisory/CVE-2018-0808
-Release Date: 2018-03-14
-Fix Resolution: Microsoft.AspNetCore.Server.IISIntegration - 2.1.0, Microsoft.AspNetCore.Hosting - 2.1.0
+ // restart node; client should resume + clus.Members[0].Restart(t) + select { + case <-donec: + case <-time.After(2 * clus.Members[1].ServerConfig.ReqTimeout()): + t.Fatalf(""waited too long"") + } + } - +``` +Core components of ASP.NET Core Kestrel cross-platform web server.
-Library home page: https://api.nuget.org/packages/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
-Path to dependency file: /TopLevel/Project1/pipelines-dotnet-core/dotnetcore-tests/dotnetcore-tests.csproj
-Path to vulnerable library: /home/wss-scanner/.nuget/packages/microsoft.aspnetcore.server.kestrel.core/2.0.0/microsoft.aspnetcore.server.kestrel.core.2.0.0.nupkg
+```go + for i := range thenOps { + clus.Members[0].Stop(t) + <-clus.Members[0].StopNotify() + + donec := make(chan struct{}, 1) + go func() { + _, err := kv.Txn(context.TODO()).Then(thenOps[i]...).Commit() + if err != nil { + t.Errorf(""expected response, got error %v"", err) + } + donec <- struct{}{} + }() + // wait for txn to fail on disconnect + time.Sleep(100 * time.Millisecond) + + // restart node; client should resume + clus.Members[0].Restart(t) + select { + case <-donec: + case <-time.After(2 * clus.Members[1].ServerConfig.ReqTimeout()): + t.Fatalf(""waited too long"") + } + } + +``` +