id
int64 393k
2.82B
| repo
stringclasses 68
values | title
stringlengths 1
936
| body
stringlengths 0
256k
⌀ | labels
stringlengths 2
508
| priority
stringclasses 3
values | severity
stringclasses 3
values |
---|---|---|---|---|---|---|
2,805,711,342 | vscode | "NVDA Not Reading When I Open a Folder from Within VSCode" |
Type: <b>Bug</b>
When I open a folder within Visual Studio Code, NVDA announces the folder as "unknown" instead of reading anything . i have to restart nvda and alt tab out and back in. This issue prevents me from knowing which folder or files are being opened, making it difficult to navigate effectively within VSCode. note, i am on a mac with a windows VM. I was also able to reproduce on my other computers.
VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z)
OS version: Windows_NT arm64 10.0.26100
Modes:
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Apple Silicon (4 x 3200)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: unavailable_off<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off|
|Load (avg)|undefined|
|Memory (System)|7.99GB (0.85GB free)|
|Process Argv|--crash-reporter-id c37eff74-c330-4ccb-910b-e92d95621bcc|
|Screen Reader|yes|
|VM|100%|
</details><details><summary>Extensions (4)</summary>
Extension|Author (truncated)|Version
---|---|---
remote-ssh|ms-|0.116.1
remote-ssh-edit|ms-|0.87.0
remote-explorer|ms-|0.4.3
pythagora-vs-code|pyt|1.2.13
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vspor879:30202332
vspor708:30202333
vspor363:30204092
vscod805:30301674
binariesv615:30325510
vsaa593:30376534
py29gd2263:31024239
c4g48928:30535728
azure-dev_surveyone:30548225
962ge761:30959799
pythonnoceb:30805159
pythonmypyd1:30879173
2e7ec940:31000449
pythontbext0:30879054
cppperfnew:31000557
dsvsc020:30976470
pythonait:31006305
dsvsc021:30996838
dvdeprecation:31068756
dwnewjupytercf:31046870
2f103344:31071589
nativerepl1:31139838
pythonrstrctxt:31112756
nativeloc1:31192215
cf971741:31144450
iacca1:31171482
notype1:31157159
5fd0e150:31155592
dwcopilot:31170013
stablechunks:31184530
6074i472:31201624
dwoutputs:31217127
hdaa2157:31222309
copilot_t_ci:31222730
```
</details>
<!-- generated by issue reporter --> | info-needed | low | Critical |
2,805,719,003 | react | Bug: Synchronously Restart Render on Error During Async Rendering to Prevent Inconsistencies | Synchronously restart when an error is thrown during async rendering
### Description
In async mode, events are interleaved with rendering. If an event mutates state that is later accessed during render, it can cause inconsistencies. Restarting the render synchronously resolves the inconsistency and prevents further mutations during interleaved events.
### Steps to Reproduce
1. Enable async rendering mode.
2. Trigger an event that mutates state during rendering.
3. Observe inconsistencies during the render process.
### Expected Behavior
The render should restart synchronously to prevent inconsistencies and further mutations from interfering.
### Actual Behavior
Inconsistent rendering occurs, leading to potential UI issues or errors.
### Suggested Fix
Implement synchronous restart of the render when an error is thrown during async rendering to handle the interleaved events and ensure consistency.
| Status: Unconfirmed,Resolution: Needs More Information | medium | Critical |
2,805,742,425 | angular | Expose `createInputSignal`, or ease restrictions on static analysis for Signal functions to reduce boilerplate | ### Which @angular/* package(s) are relevant/related to the feature request?
core
### Description
The Angular CDK provides `coercion` functions to make it easy to transform `input` values. For example:
```ts
import { BooleanInput, coerceBooleanProperty } from '@angular/cdk/coercion';
public readonly canEdit = input<boolean, BooleanInput>(false, { transform: coerceBooleanProperty });
public readonly canDelete = input<boolean, BooleanInput>(false, { transform: coerceBooleanProperty });
```
This makes it easy to use the component like below, where simply the presence of an empty `canEdit` or `canDelete` attribute will set the `input` Signal's value to `true`:
```html
<my-component canEdit canDelete />
```
When using something like Prettier with an 80-char line length, each input defined like this becomes three lines long:
```ts
public readonly canEdit = input(false, {
transform: coerceBooleanProperty,
});
public readonly canDelete = input(true, {
transform: coerceBooleanProperty,
});
```
This is a lot of boilerplate. Ideally, I'd be able to do create a utility function like the following:
```ts
export function booleanInput(defaultValue: boolean): InputSignalWithTransform<boolean, BooleanInput> {
return input(defaultValue, { transform: coerceBooleanProperty });
}
```
```ts
public readonly canEdit = booleanInput(false);
public readonly canDelete = booleanInput(false);
```
However the above code isn't valid, because:
`✘ [ERROR] TS-998110: Unsupported call to the input function. This function can only be called in the initializer of a class member. [plugin angular-compiler]`
That's fair, so next I figured I'd try and simplify things by doing:
```ts
const boolTransform: InputOptionsWithTransform<boolean, BooleanInput> = {
transform: coerceBooleanProperty
};
```
```ts
public readonly canEdit = input(false, boolTransform);
public readonly canDelete = input(false, boolTransform);
```
However this **also** isn't valid, because:
`✘ [ERROR] TS-991010: Argument needs to be an object literal that is statically analyzable. [plugin angular-compiler]`
I understand why both of these restrictions exist, but it would be nice to be able to avoid needing to rewrite out the entire options object for every Signal when they are all exactly the same.
### Proposed solution
I'm unsure of any knock-on effects of exposing it, but `core/src/authoring/input` contains the below function:
```ts
export function createInputSignal<T, TransformT>(
initialValue: T,
options?: InputOptions<T, TransformT>,
): InputSignalWithTransform<T, TransformT>;
```
If this were exposed from `@angular/core`, it could be possible to write wrapper utilities for signals such as the `booleanInput()` example I described, or other shortcuts such as a number-based signal that has min/max values and normalizes values to be within range, etc.
### Alternatives considered
Alternatively, relaxing (or expanding) the static analysis of Options objects would mean that it would be easier to repeat the same transforms/options throughout the codebase even when using the base `input`/`model`/etc functions. | area: compiler,cross-cutting: signals | low | Critical |
2,805,760,863 | ollama | API parameter: 'reasoning_effort' (for DeepSeek-R1) | When I will be able to use `reasoning_effort` like on DeepSeek-R1? :)
Doc: https://api-docs.deepseek.com/guides/reasoning_model | feature request | low | Minor |
2,805,761,049 | react | Bug: Prevent event log from growing unbounded | When a Scheduler profile runs without stopping, the event log will grow unbounded. Eventually it will run out of memory and the VM will throw an error.
React version: [v16.9.0](https://github.com/facebook/react/releases/tag/v16.9.0)
## Steps To Reproduce
1. Run Scheduler profile
2. Watch memory resource growing
## The current behavior
When Scheduler profile starts it runs out of memory and the JVM throws an error
## The expected behavior
The event log should not grow unbounded | Status: Unconfirmed | medium | Critical |
2,805,765,889 | flutter | [go_router] [go_router_builder] go_router already supports preloading but go_router_builder doesn't | ### Use case
Support preload for `StatefulShellBranch` when use `go_router_builder`
### Proposal
Add `preload` field to `go_router_builder` and code gen helper function of `go_router` | c: new feature,package,c: proposal,p: go_router_builder,team-go_router | low | Minor |
2,805,773,579 | pytorch | [Dynamo]while_loop raise an exception | ### 🐛 Describe the bug
while loop will raise an exception.
The reason is input tensor has been sliced.
```
tensor strided info: shape: torch.Size([192]), original_tshape: [64], slice:slice(None, None, 3)
```
This creates tensor with stride=3, but once the body_fn execute tensor+1 the stride is reset to 1 and this causes the error.
If tensor which is strided by nature due to slice before passing to while_loop then pytorch doesn't accept this.
mini reproducer:
```
import torch
class OpWrapperModule(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, ifm, op_inputs_dict):
result = torch.while_loop(**op_inputs_dict)
return result
torch.manual_seed(7616)
ifm_t = torch.randn([64])
ifm = ifm_t[slice(None, None, 3)]
iterations = torch.tensor(50)
def cond_fn(inputs, iterations):
return iterations > 0
def body_fn(inputs, iterations):
return [inputs + 1, iterations - 1]
params = {
"cond_fn": cond_fn,
"body_fn": body_fn,
"carried_inputs": (ifm, iterations),
}
model = OpWrapperModule()
result = model(ifm, params)
```
ERROR log and trace:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/variables/higher_order_ops.py", line 55, in graph_break_as_hard_error
return fn(*args, **kwargs)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/variables/higher_order_ops.py", line 1115, in call_function
unimplemented(
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/exc.py", line 317, in unimplemented
raise Unsupported(msg, case_name=case_name)
torch._dynamo.exc.Unsupported: Expected carried_inputs and body outputs return tensors with same metadata but find:
pair[0] differ in 'stride: (3,) vs (1,)', where lhs is TensorMetadata(shape=torch.Size([22]), dtype=torch.float32, requires_grad=False, stride=(3,), memory_format=None, is_quantized=False, qparams={}) and rhs is TensorMetadata(shape=torch.Size([22]), dtype=torch.float32, requires_grad=False, stride=(1,), memory_format=None, is_quantized=False, qparams={})
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/zhenzhao/qnpu/dynamo/src/pytorch-training-tests/tests/torch_feature_val/single_op/repro_while_loop.py", line 28, in <module>
result = model(ifm, params)
File "/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py", line 1742, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py", line 1753, in _call_impl
return forward_call(*args, **kwargs)
File "/home/zhenzhao/qnpu/dynamo/src/pytorch-training-tests/tests/torch_feature_val/single_op/repro_while_loop.py", line 8, in forward
result = torch.while_loop(**op_inputs_dict)
File "/usr/local/lib/python3.10/dist-packages/torch/_higher_order_ops/while_loop.py", line 165, in while_loop
return torch.compile(
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/eval_frame.py", line 574, in _fn
return fn(*args, **kwargs)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 1380, in __call__
return self._torchdynamo_orig_callable(
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 547, in __call__
return _compile(
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 986, in _compile
guarded_code = compile_inner(code, one_graph, hooks, transform)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 715, in compile_inner
return _compile_inner(code, one_graph, hooks, transform)
File "/usr/local/lib/python3.10/dist-packages/torch/_utils_internal.py", line 95, in wrapper_function
return function(*args, **kwargs)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 750, in _compile_inner
out_code = transform_code_object(code, transform)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/bytecode_transformation.py", line 1361, in transform_code_object
transformations(instructions, code_options)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 231, in _fn
return fn(*args, **kwargs)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 662, in transform
tracer.run()
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 2868, in run
super().run()
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 1052, in run
while self.step():
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 962, in step
self.dispatch_table[inst.opcode](self, inst)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 659, in wrapper
return inner_fn(self, inst)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 1736, in CALL_FUNCTION_EX
self.call_function(fn, argsvars.items, kwargsvars)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 897, in call_function
self.push(fn.call_function(self, args, kwargs)) # type: ignore[arg-type]
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/variables/higher_order_ops.py", line 58, in graph_break_as_hard_error
raise UncapturedHigherOrderOpError(reason + msg) from e
torch._dynamo.exc.UncapturedHigherOrderOpError: while_loop doesn't work unless it is captured completely with torch.compile. Scroll up to find out what causes the graph break.
from user code:
File "/usr/local/lib/python3.10/dist-packages/torch/_higher_order_ops/while_loop.py", line 156, in _while_loop_op_wrapper
return while_loop_op(*args, **kwargs)
### Versions
PyTorch version: 2.6.0a0+git30ac7fd
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
Clang version: 14.0.5 (ssh://[[email protected]](mailto:[email protected])/habana-internal/tpc_llvm10 150d2d7c6a8ff8abf0d8ce194d3fac3986b078e6)
CMake version: version 3.28.1
Libc version: glibc-2.35
Python version: 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0] (64-bit runtime)
Python platform: Linux-5.15.0-127-generic-x86_64-with-glibc2.35
Is CUDA available: False
CUDA runtime version: No CUDA
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 43 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 12
On-line CPU(s) list: 0-11
Vendor ID: GenuineIntel
Model name: Intel(R) Xeon(R) Gold 6132 CPU @ 2.60GHz
CPU family: 6
Model: 85
Thread(s) per core: 1
Core(s) per socket: 6
Socket(s): 2
Stepping: 0
BogoMIPS: 5187.81
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon nopl xtopology tsc_reliable nonstop_tsc cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch invpcid_single pti ssbd ibrs ibpb stibp tpr_shadow vnmi ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 invpcid avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xsaves arat pku ospke md_clear flush_l1d arch_capabilities
Virtualization: VT-x
Hypervisor vendor: VMware
Virtualization type: full
L1d cache: 384 KiB (12 instances)
L1i cache: 384 KiB (12 instances)
L2 cache: 12 MiB (12 instances)
L3 cache: 38.5 MiB (2 instances)
NUMA node(s): 1
NUMA node0 CPU(s): 0-11
Vulnerability Gather data sampling: Unknown: Dependent on hypervisor status
Vulnerability Itlb multihit: KVM: Mitigation: VMX disabled
Vulnerability L1tf: Mitigation; PTE Inversion; VMX flush not necessary, SMT disabled
Vulnerability Mds: Mitigation; Clear CPU buffers; SMT Host state unknown
Vulnerability Meltdown: Mitigation; PTI
Vulnerability Mmio stale data: Mitigation; Clear CPU buffers; SMT Host state unknown
Vulnerability Reg file data sampling: Not affected
Vulnerability Retbleed: Mitigation; IBRS
Vulnerability Spec rstack overflow: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; IBRS; IBPB conditional; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI SW loop, KVM SW loop
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
cc @chauhang @penguinwu @zou3519 @ydwu4 @bdhirsh @yf225 | triaged,oncall: pt2,module: higher order operators,module: pt2-dispatcher | low | Critical |
2,805,777,333 | PowerToys | Keeps forgetting that I have fancy zones turned on. | ### Microsoft PowerToys version
0.87.1
### Installation method
Other (please specify in "Steps to Reproduce")
### Running as admin
Yes
### Area(s) with issue?
FancyZones
### Steps to reproduce
Installation method don't remember. From Google search.
It randomly forgets that I have fancy zones turned on (great feature, b.t.w. :) ), and I have to manually turn it on again.
### ✔️ Expected Behavior
It remembers that I have fancy zones turned on, for each user.
### ❌ Actual Behavior
It randomly forgets.
### Other Software
_No response_ | Issue-Bug,Needs-Triage | low | Minor |
2,805,797,414 | react | Bug: Fix Buffer - Write Chunks to Separate Buffers | <!--
Please provide a clear and concise description of what the bug is. Include
screenshots if needed. Please test using the latest version of the relevant
React packages to make sure your issue has not already been fixed.
-->
The bug happens when using ReactDOMFizzServer.renderToReadableStream to render big chunks of data. If a chunk is bigger than the buffer (default: 512 bytes), the buffer overflows or uses incomplete data. This causes wrong or cut-off output because the renderer doesn’t handle buffer size properly.
React version:
React 19.0.0
## Steps To Reproduce
1. Create a large chunk of data that is bigger than the buffer size.
2. Start processing the data using `ReactDOMFizzServer.renderToReadableStream`.
3. Observe the behavior when the chunks exceed the buffer size.
<!--
Your bug will get fixed much faster if we can run your code and it doesn't
have dependencies other than React. Issues without reproduction steps or
code examples may be immediately closed as not actionable.
-->
<!--
Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a
repository on GitHub, or provide a minimal code example that reproduces the
problem. You may provide a screenshot of the application if you think it is
relevant to your bug report. Here are some tips for providing a minimal
example: https://stackoverflow.com/help/mcve.
-->
```
import ReactDOMFizzServer from 'react-dom/server';
// Minimal data to reproduce the problem
const largeChunk = 'A'.repeat(2049); // A chunk exceeding default 512 bytes
async function testStream() {
// simple React structure with the large chunk
const stream = await ReactDOMFizzServer.renderToReadableStream(
<div>{largeChunk}</div>
);
// function to read the stream
const decoder = new TextDecoder();
const reader = stream.getReader();
let result = '';
while (true) {
const { done, value } = await reader.read();
if (done) break;
result += decoder.decode(value);
}
console.log(result);
}
testStream();
```
## The current behavior
Right now, when chunks are sent to a `ReadableStream`, they are enqueued directly without checking the buffer size. If a chunk is too big (like 2049 bytes, while the buffer is only 512 bytes), it reuses old buffers, which messes up the output. For example, instead of properly rendering, we get something like ```<div>AAAAA...</div>```.
## The expected behavior
Chunks should be written to their own buffer and split into smaller parts if they are too big. The buffer should only be sent once it's full or the writing is complete. This way, even large chunks (like 2049 bytes) will render correctly, for example: ```<div>AAAAAAAAAAAAAAAAA...</div>```.
| Status: Unconfirmed | medium | Critical |
2,805,814,674 | pytorch | [CUDA] Illegal Memory Access with `AdaptiveMaxPool2d` | ### 🐛 Describe the bug
Found by fuzzer, the parameter used to trigger this does not look too corner-case.
@eqy
```python
import torch
m1 = torch.randn(8812, 1, 2).cuda()
model = torch.nn.AdaptiveMaxPool2d(output_size=[262143, 1]).cuda()
model(m1)
```
```bash
computer-sanitizer python3 poc7.py
```
compute-sanitizer log.
```python
========= Invalid __global__ write of size 8 bytes
========= at void at::native::<unnamed>::adaptivemaxpool<float>(const T1 *, T1 *, long *, int, int, int, int, long, long, long)+0x1b80
========= by thread (0,4,0) in block (8195,0,0)
========= Address 0x79d8025f0008 is out of bounds
========= and is 17,173,643,256 bytes before the nearest allocation at 0x79dc02000000 of size 18,480,103,424 bytes
========= Saved host backtrace up to driver entry point at kernel launch time
========= Host Frame: [0x2dfbef]
========= in /lib/x86_64-linux-gnu/libcuda.so.1
========= Host Frame: [0x15803]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/../../../../libcudart.so.12
========= Host Frame:cudaLaunchKernel [0x75230]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/../../../../libcudart.so.12
========= Host Frame:at::native::structured_adaptive_max_pool2d_out_cuda::impl(at::Tensor const&, c10::ArrayRef<long>, at::Tensor const&, at::Tensor const&)::{lambda()#1}::operator()() const [0x160853b]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so
========= Host Frame:at::native::structured_adaptive_max_pool2d_out_cuda::impl(at::Tensor const&, c10::ArrayRef<long>, at::Tensor const&, at::Tensor const&) [0x160a8f4]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so
========= Host Frame:at::(anonymous namespace)::wrapper_CUDA_adaptive_max_pool2d(at::Tensor const&, c10::ArrayRef<long>) [0x3606f1e]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so
========= Host Frame:c10::impl::wrap_kernel_functor_unboxed_<c10::impl::detail::WrapFunctionIntoFunctor_<c10::CompileTimeFunctionPointer<std::tuple<at::Tensor, at::Tensor> (at::Tensor const&, c10::ArrayRef<long>), &at::(anonymous namespace)::wrapper_CUDA_adaptive_max_p
ool2d>, std::tuple<at::Tensor, at::Tensor>, c10::guts::typelist::typelist<at::Tensor const&, c10::ArrayRef<long> > >, std::tuple<at::Tensor, at::Tensor> (at::Tensor const&, c10::ArrayRef<long>)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, c10::Array
Ref<long>) [0x3606fe2]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so
========= Host Frame:at::_ops::adaptive_max_pool2d::redispatch(c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>) [0x27406bd]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cpu.so
========= Host Frame:torch::autograd::VariableType::(anonymous namespace)::adaptive_max_pool2d(c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>) [0x49ee7d4]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cpu.so
========= Host Frame:c10::impl::wrap_kernel_functor_unboxed_<c10::impl::detail::WrapFunctionIntoFunctor_<c10::CompileTimeFunctionPointer<std::tuple<at::Tensor, at::Tensor> (c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>), &torch::autograd::VariableType::(a
nonymous namespace)::adaptive_max_pool2d>, std::tuple<at::Tensor, at::Tensor>, c10::guts::typelist::typelist<c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long> > >, std::tuple<at::Tensor, at::Tensor> (c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>)>::
call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, c10::ArrayRef<long>) [0x49eef95]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cpu.so
========= Host Frame:at::_ops::adaptive_max_pool2d::call(at::Tensor const&, c10::ArrayRef<long>) [0x27cb9db]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_cpu.so
========= Host Frame:torch::autograd::THPVariable_adaptive_max_pool2d(_object*, _object*, _object*) [0x765397]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/lib/python3.12/site-packages/torch/lib/libtorch_python.so
========= Host Frame:cfunction_call in /usr/local/src/conda/python-3.12.7/Objects/methodobject.c:537 [0x149d53]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyObject_MakeTpCall in /usr/local/src/conda/python-3.12.7/Objects/call.c:240 [0x11af9a]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyEval_EvalFrameDefault in Python/bytecodes.c:2715 [0x125902]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:method_vectorcall in /usr/local/src/conda/python-3.12.7/Objects/classobject.c:91 [0x175716]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyEval_EvalFrameDefault in Python/bytecodes.c:3263 [0x12acac]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:method_vectorcall in /usr/local/src/conda/python-3.12.7/Objects/classobject.c:91 [0x175716]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyEval_EvalFrameDefault in Python/bytecodes.c:3263 [0x12acac]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyObject_FastCallDictTstate in /usr/local/src/conda/python-3.12.7/Objects/call.c:133 [0x11db06]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyObject_Call_Prepend in /usr/local/src/conda/python-3.12.7/Objects/call.c:508 [0x157e55]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:slot_tp_call in /usr/local/src/conda/python-3.12.7/Objects/typeobject.c:8782 [0x22f065]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyObject_MakeTpCall in /usr/local/src/conda/python-3.12.7/Objects/call.c:240 [0x11af9a]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyEval_EvalFrameDefault in Python/bytecodes.c:2715 [0x125902]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:PyEval_EvalCode in /usr/local/src/conda/python-3.12.7/Python/ceval.c:578 [0x1e3c6d]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:run_eval_code_obj in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:1722 [0x20a0b6]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:run_mod in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:1743 [0x2056d6]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:pyrun_file in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:1643 [0x21d601]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyRun_SimpleFileObject in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:433 [0x21cf3f]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:_PyRun_AnyFileObject in /usr/local/src/conda/python-3.12.7/Python/pythonrun.c:78 [0x21cd32]
========= in /home/jwnhy/miniconda3/envs/gpu-torch/bin/python3
========= Host Frame:Py_RunMain in /usr/local/src/conda/python-3.12.7/Modules/main.c:713 [0x215dc2]
[
```
### Versions
```
Collecting environment information...
PyTorch version: 2.5.1
Is debug build: False
CUDA used to build PyTorch: 12.4
ROCM used to build PyTorch: N/A
OS: Ubuntu 24.04.1 LTS (x86_64)
GCC version: (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0
Clang version: Could not collect
CMake version: Could not collect
Libc version: glibc-2.39
Python version: 3.12.7 | packaged by Anaconda, Inc. | (main, Oct 4 2024, 13:27:36) [GCC 11.2.0] (64-bit runtime)
Python platform: Linux-6.11.0-1007-oem-x86_64-with-glibc2.39
Is CUDA available: True
CUDA runtime version: 12.6.85
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration: GPU 0: NVIDIA H100 PCIe
Nvidia driver version: 560.35.05
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 46 bits physical, 57 bits virtual
Byte Order: Little Endian
CPU(s): 48
On-line CPU(s) list: 0-47
Vendor ID: GenuineIntel
Model name: INTEL(R) XEON(R) SILVER 4510
CPU family: 6
Model: 143
Thread(s) per core: 2
Core(s) per socket: 12
Socket(s): 2
Stepping: 8
CPU(s) scaling MHz: 37%
CPU max MHz: 4100.0000
CPU min MHz: 800.0000
BogoMIPS: 4800.00
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust sgx bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect user_shstk avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts vnmi avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd sgx_lc fsrm md_clear serialize tsxldtrk pconfig arch_lbr ibt amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities
Virtualization: VT-x
L1d cache: 1.1 MiB (24 instances)
L1i cache: 768 KiB (24 instances)
L2 cache: 48 MiB (24 instances)
L3 cache: 60 MiB (2 instances)
NUMA node(s): 2
NUMA node0 CPU(s): 0-11,24-35
NUMA node1 CPU(s): 12-23,36-47
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Reg file data sampling: Not affected
Vulnerability Retbleed: Not affected
Vulnerability Spec rstack overflow: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS SW sequence; BHI BHI_DIS_S
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] numpy==2.1.3
[pip3] nvidia-cublas-cu12==12.6.4.1
[pip3] nvidia-cudnn-cu12==9.5.1.17
[pip3] torch==2.5.1
[pip3] torchaudio==2.5.1
[pip3] torchvision==0.20.1
[pip3] triton==3.1.0
[conda] blas 1.0 mkl
[conda] cuda-cudart 12.4.127 0 nvidia
[conda] cuda-cupti 12.4.127 0 nvidia
[conda] cuda-libraries 12.4.1 0 nvidia
[conda] cuda-nvrtc 12.4.127 0 nvidia
[conda] cuda-nvtx 12.4.127 0 nvidia
[conda] cuda-opencl 12.6.77 0 nvidia
[conda] cuda-runtime 12.4.1 0 nvidia
[conda] ffmpeg 4.3 hf484d3e_0 pytorch
[conda] libcublas 12.4.5.8 0 nvidia
[conda] libcufft 11.2.1.3 0 nvidia
[conda] libcurand 10.3.7.77 0 nvidia
[conda] libcusolver 11.6.1.9 0 nvidia
[conda] libcusparse 12.3.1.170 0 nvidia
[conda] libjpeg-turbo 2.0.0 h9bf148f_0 pytorch
[conda] libnvjitlink 12.4.127 0 nvidia
[conda] mkl 2023.1.0 h213fc3f_46344
[conda] mkl-service 2.4.0 py312h5eee18b_1
[conda] mkl_fft 1.3.11 py312h5eee18b_0
[conda] mkl_random 1.2.8 py312h526ad5a_0
[conda] numpy 2.1.3 py312hc5e2394_0
[conda] numpy-base 2.1.3 py312h0da6c21_0
[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi
[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi
[conda] pytorch 2.5.1 py3.12_cuda12.4_cudnn9.1.0_0 pytorch
[conda] pytorch-cuda 12.4 hc786d27_7 pytorch
[conda] pytorch-mutex 1.0 cuda pytorch
[conda] torchaudio 2.5.1 py312_cu124 pytorch
[conda] torchtriton 3.1.0 py312 pytorch
[conda] torchvision 0.20.1 py312_cu124 pytorch
```
cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki @ptrblck @msaroufim @eqy | module: nn,module: cuda,triaged,module: edge cases | low | Critical |
2,805,822,844 | pytorch | torch._dynamo.exc.Unsupported: Graph break due to unsupported builtin torch._C._dynamo.eval_frame.set_eval_frame. | ### 🐛 Describe the bug
Part of an example as the following:
```python
from torch_mlir import fx
module = fx.export_and_import(
model,
input_tensor,
output_type="tosa",
func_name="forward",
)
```
when I use pyinstaller to pack the fx graph export and import to a binary file, it encounters the following problem. Actually, when I run the unpacked python code, it works well.

### Versions
`torch==2.6.0.dev20241216+cpu`
cc @chauhang @penguinwu @avikchaudhuri @gmagogsfm @zhxchen17 @tugsbayasgalan @angelayi @suo @ydwu4 | oncall: pt2,export-triaged,oncall: export | low | Critical |
2,805,827,334 | rust | Error installing cargo-watch | <!--
Thank you for finding an Internal Compiler Error! 🧊 If possible, try to provide
a minimal verifiable example. You can read "Rust Bug Minimization Patterns" for
how to create smaller examples.
http://blog.pnkfx.org/blog/2019/11/18/rust-bug-minimization-patterns/
-->
### Code
```Rust
cargo install cargo-watch
```
### Meta
`rustc --version --verbose`:
```
rustc 1.84.0 (9fc6b4312 2025-01-07)
binary: rustc
commit-hash: 9fc6b43126469e3858e2fe86cafb4f0fd5068869
commit-date: 2025-01-07
host: x86_64-unknown-linux-gnu
release: 1.84.0
LLVM version: 19.1.5
```
### Error output
```
error: the compiler unexpectedly panicked. this is a bug.
note: we would appreciate a bug report: https://github.com/rust-lang/rust/issues/new?labels=C-bug%2C+I-ICE%2C+T-compiler&template=ice.md
note: rustc 1.84.0 (9fc6b4312 2025-01-07) running on x86_64-unknown-linux-gnu
note: compiler flags: --crate-type lib -C opt-level=3 -C panic=abort -C linker-plugin-lto -C codegen-units=1 -C strip=debuginfo
note: some of the compiler flags provided by cargo are hidden
```
```
thread 'main' panicked at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/std/src/thread/mod.rs:1719:40:
called `Option::unwrap()` on a `None` value
stack backtrace:
Compiling inotify v0.7.1
0: 0x7148abd8682a - <std::sys::backtrace::BacktraceLock::print::DisplayBacktrace as core::fmt::Display>::fmt::ha4a311b32f6b4ad8
1: 0x7148ac4277e2 - core::fmt::write::h1866771663f62b81
2: 0x7148ad281d51 - std::io::Write::write_fmt::hb549e7444823135e
3: 0x7148abd86682 - std::sys::backtrace::BacktraceLock::print::hddd3a9918ce29aa7
4: 0x7148abd88b5a - std::panicking::default_hook::{{closure}}::h791f75256b902d7d
5: 0x7148abd889c0 - std::panicking::default_hook::h82cc572fcb0d8cd7
6: 0x7148aae14f55 - std[1b49f43dde054edc]::panicking::update_hook::<alloc[f0e0d4128a1437e6]::boxed::Box<rustc_driver_impl[c421ed190efad9be]::install_ice_hook::{closure#0}>>::{closure#0}
7: 0x7148abd89238 - std::panicking::rust_panic_with_hook::he21644cc2707f2c4
8: 0x7148abd88fd6 - std::panicking::begin_panic_handler::{{closure}}::h42f7c414fed3cad9
9: 0x7148abd86cd9 - std::sys::backtrace::__rust_end_short_backtrace::ha26cf5766b4e8c65
10: 0x7148abd88ccc - rust_begin_unwind
11: 0x7148a88e43f0 - core::panicking::panic_fmt::h74866b78e934b1c0
12: 0x7148a8afe56c - core::panicking::panic::h95d8269cf8bd4f7a
13: 0x7148a9e65909 - core::option::unwrap_failed::hca433a9693b91bd2
14: 0x7148ad27a0d9 - rustc_driver_impl[c421ed190efad9be]::run_compiler
15: 0x7148ad1c99e0 - rustc_driver_impl[c421ed190efad9be]::main
16: 0x5eadf5e9e497 - rustc_main[302bc4b73869fe48]::main
17: 0x5eadf5e9e483 - std[1b49f43dde054edc]::sys::backtrace::__rust_begin_short_backtrace::<fn(), ()>
18: 0x5eadf5e9e479 - std[1b49f43dde054edc]::rt::lang_start::<()>::{closure#0}
19: 0x7148ad29dd41 - std::rt::lang_start_internal::h78dd36c15a6b42b8
20: 0x5eadf5ea5ea7 - main
21: 0x7148a7429d90 - <unknown>
22: 0x7148a7429e40 - __libc_start_main
23: 0x5eadf5ea5db6 - <unknown>
24: 0x0 - <unknown>
Compiling clap v2.34.0
```
</details>
| I-ICE,T-compiler,C-bug,S-needs-repro | low | Critical |
2,805,829,310 | pytorch | DISABLED test_tensor_subclass_basic (__main__.TestCompiledAutograd) | Platforms: asan, linux, rocm, mac, macos, slow
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_tensor_subclass_basic&suite=TestCompiledAutograd&limit=100) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/36028557832).
Over the past 3 hours, it has been determined flaky in 20 workflow(s) with 40 failures and 20 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_tensor_subclass_basic`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `inductor/test_compiled_autograd.py`
cc @clee2000 @wdvr @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @chauhang @aakhundov | triaged,module: flaky-tests,skipped,oncall: pt2,module: inductor | low | Critical |
2,805,848,501 | electron | I am currently upgrading the project from Electron 15 to 29, but the embedded document websites consistently pop up a font download popup window | ### Preflight Checklist
- [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project.
- [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [x] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success.
### Electron Version
29.4.26
### What operating system(s) are you using?
macOS
### Operating System Version
13.2.1 (22D68)
### What arch are you using?
arm64 (including Apple Silicon)
### Last Known Working Electron version
30.0.0
### Expected Behavior
I hope that, just like in Electron 15, this pop - up can be skipped entirely, and the default font can be used directly. This is because such an experience is really bad for users.
### Actual Behavior
Request for Support Background: I am currently upgrading the project from Electron 15 to Electron 29, but the embedded document websites consistently pop up a font download popup window. Problem: When using BrowserView, WebView, or iframe in the project to reference external websites, this popup window will appear when a specific font does not exist on Mac. This font download popup window did not exist in version 15. I want to keep it consistent with version 15. Which feature of Electron or Chrome has led to this phenomenon? How should I set it? Below is a screenshot.
Additionally, through my tests in fiddle, I found that this issue still persists in the last version 29.4.26 of Electron 29. However, this problem does not occur in version 30.0.0. Also, there is no such problem in the last version of Electron 28

### Testcase Gist URL
https://gist.github.com/Tsite2007/aee74ff4afe6f14148d88a1aacb74364
### Additional Information
_No response_ | platform/macOS,blocked/need-info ❌,bug :beetle:,has-repro-gist | low | Critical |
2,805,853,403 | flutter | Engine `local_engine_builds` failing since #161825 | `local_engine_builds` has been in bringup, which means https://github.com/flutter/flutter/pull/161825 broke it without knowing.
(https://github.com/flutter/flutter/pull/160627 initially marked it as `bringup: true`, @jonahwilliams was this intended to be permanent?)
@eyebrowsoffire can you fix this (or revert your changes if you don't imagine this happening for some time)? | engine,P1,c: tech-debt,team-engine,triaged-engine,fyi-web,e: engine-tool | medium | Minor |
2,805,876,375 | vscode | Terminal font ligatures doesn't render as expected | In terminal

In editor

Version: 1.97.0-insider (user setup)
Commit: d226a2a497b928d78aa654f74c8af5317d3becfb
Date: 2025-01-22T05:05:14.565Z
Electron: 32.2.7
ElectronBuildId: 10660205
Chromium: 128.0.6613.186
Node.js: 20.18.1
V8: 12.8.374.38-electron.0
OS: Windows_NT x64 10.0.22621 | info-needed,terminal-ligatures | low | Major |
2,805,889,854 | pytorch | Flex Attention not support score_mod with gradients | ### 🐛 Describe the bug
`Flex Attention` does not support `score_mod with gradients`, making it impossible to define a learnable score_mod for the dynamic mask attention variants.
```python
import torch
from torch.nn.attention.flex_attention import flex_attention
A = torch.nn.Parameter(torch.zeros(1))
query = torch.randn(1, 1, 1, 32)
key = torch.randn(1, 1, 1, 32)
value = torch.randn(1, 1, 1, 32)
dt = torch.randn(1, 1, 1)
dynamic_mask = torch.exp(A * dt).transpose(-1, -2)
attn_mask = dynamic_mask[:, :, None, :]
def dynamic_mod(score, batch, head, q_idx, kv_idx):
score = score + attn_mask[batch][head][0][kv_idx]
return score
attn_output = flex_attention(
query=query,
key=key,
value=value,
score_mod=dynamic_mod,
)
print(attn_output)
```
```
Traceback (most recent call last):
File "e:\Doge\bug.py", line 22, in <module>
attn_output = flex_attention(
File "E:\conda\envs\doge\lib\site-packages\torch\nn\attention\flex_attention.py", line 1038, in flex_attention
out, lse = torch.compile(
File "E:\conda\envs\doge\lib\site-packages\torch\_dynamo\eval_frame.py", line 465, in _fn
return fn(*args, **kwargs)
File "E:\conda\envs\doge\lib\site-packages\torch\nn\attention\flex_attention.py", line 1032, in _flex_attention_hop_wrapper
def _flex_attention_hop_wrapper(*args, **kwargs):
File "E:\conda\envs\doge\lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
return fn(*args, **kwargs)
File "<eval_with_key>.3", line 24, in forward
File "E:\conda\envs\doge\lib\site-packages\torch\_higher_order_ops\flex_attention.py", line 109, in __call__
return super().__call__(
File "E:\conda\envs\doge\lib\site-packages\torch\_ops.py", line 433, in __call__
return wrapper()
File "E:\conda\envs\doge\lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
return fn(*args, **kwargs)
File "E:\conda\envs\doge\lib\site-packages\torch\_ops.py", line 429, in wrapper
return self.dispatch(
File "E:\conda\envs\doge\lib\site-packages\torch\_ops.py", line 412, in dispatch
return kernel(*args, **kwargs)
File "E:\conda\envs\doge\lib\site-packages\torch\_higher_order_ops\flex_attention.py", line 703, in flex_attention_autograd
out, logsumexp = FlexAttentionAutogradOp.apply(
File "E:\conda\envs\doge\lib\site-packages\torch\autograd\function.py", line 575, in apply
return super().apply(*args, **kwargs) # type: ignore[misc]
File "E:\conda\envs\doge\lib\site-packages\torch\_higher_order_ops\flex_attention.py", line 578, in forward
assert (
AssertionError: Captured buffers that require grad are not yet supported.
```
### Versions
Collecting environment information...
PyTorch version: 2.6.0a0+df5bbc09d1.nv24.12
Is debug build: False
CUDA used to build PyTorch: 12.6
ROCM used to build PyTorch: N/A
OS: Ubuntu 24.04.1 LTS (x86_64)
GCC version: (Ubuntu 13.2.0-23ubuntu4) 13.2.0
Clang version: Could not collect
CMake version: version 3.31.1
Libc version: glibc-2.39
Python version: 3.12.3 (main, Nov 6 2024, 18:32:19) [GCC 13.2.0] (64-bit runtime)
Python platform: Linux-5.15.153.1-microsoft-standard-WSL2-x86_64-with-glibc2.39
Is CUDA available: True
CUDA runtime version: 12.6.85
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration: GPU 0: NVIDIA GeForce RTX 4090
Nvidia driver version: 566.36
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.9.6.0
/usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.6.0
/usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.6.0
/usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.6.0
/usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.6.0
/usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.6.0
/usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.6.0
/usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.6.0
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 46 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 32
On-line CPU(s) list: 0-31
Vendor ID: GenuineIntel
Model name: 13th Gen Intel(R) Core(TM) i9-13900KS
CPU family: 6
Model: 183
Thread(s) per core: 2
Core(s) per socket: 16
Socket(s): 1
Stepping: 1
BogoMIPS: 6374.42
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology tsc_reliable nonstop_tsc cpuid pni pclmulqdq vmx ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves avx_vnni umip waitpkg gfni vaes vpclmulqdq rdpid movdiri movdir64b fsrm md_clear serialize flush_l1d arch_capabilities
Virtualization: VT-x
Hypervisor vendor: Microsoft
Virtualization type: full
L1d cache: 768 KiB (16 instances)
L1i cache: 512 KiB (16 instances)
L2 cache: 32 MiB (16 instances)
L3 cache: 36 MiB (1 instance)
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Retbleed: Mitigation; Enhanced IBRS
Vulnerability Spec rstack overflow: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Enhanced IBRS, IBPB conditional, RSB filling, PBRSB-eIBRS SW sequence
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] mypy-extensions==1.0.0
[pip3] numpy==1.26.4
[pip3] nvidia-cudnn-frontend==1.8.0
[pip3] nvtx==0.2.5
[pip3] onnx==1.17.0
[pip3] optree==0.13.1
[pip3] pynvjitlink==0.3.0
[pip3] pytorch-triton==3.0.0+72734f086
[pip3] torch==2.6.0a0+df5bbc09d1.nv24.12
[pip3] torch_tensorrt==2.6.0a0
[pip3] torchprofile==0.0.4
[pip3] torchvision==0.20.0a0
[conda] Could not collect
cc @chauhang @penguinwu @zou3519 @ydwu4 @bdhirsh @yf225 @Chillee @drisspg @yanboliang @BoyuanFeng | triaged,oncall: pt2,module: higher order operators,module: pt2-dispatcher,module: flex attention | low | Critical |
2,805,914,111 | godot | (4.4.beta1) Android app crashes when exported with .NET 9 | ### Tested versions
**Reproducible:**
4.4.beta1
4.4.dev7
**Not Reproducible:**
4.4.dev3
### System information
Godot v4.4.beta1.mono - Windows 10 (build 19045) - OpenGL 3 (Compatibility)
### Issue description
When exporting to Android with .NET 9 using Gradle Build via the Remote Deploy, the app crashes immediately at startup.
Note: .NET 8 works fine, but .NET 9 causes the app to crash
using the command at runtime:
**adb logcat|grep godot**
> 01-22 19:27:18.818 1895 3450 I ActivityTaskManager: START u0 {act=android.intent.action.MAIN flg=0x10000000 cmp=com.example.test_android/com.godot.game.GodotApp} from uid 2000
01-22 19:27:18.857 1895 2224 I ActivityManager: Start proc 1876:com.example.test_android/u0a411 for next-top-activity {com.example.test_android/com.godot.game.GodotApp}
01-22 19:27:19.885 1895 2006 W ActivityTaskManager: Force finishing activity com.example.test_android/com.godot.game.GodotApp
01-22 19:27:19.987 1895 2220 I GameServiceProviderInstance: Failed to remove task overlay. This is expected if the task is already destroyed: GameSessionRecord{mTaskId=306, mState=GAME_SESSION_ATTACHED, mRootComponentName=ComponentInfo{com.example.test_android/com.godot.game.GodotApp}, mIGameSession=android.service.games.IGameSession$Stub$Proxy@3a45eaa, mSurfacePackage=android.view.SurfaceControlViewHost$SurfacePackage@3504a9b}
01-22 19:27:20.386 1895 2200 W ActivityTaskManager: Activity top resumed state loss timeout for ActivityRecord{691de90 u0 com.example.test_android/com.godot.game.GodotApp} t-1 f}}
### Steps to reproduce
1. Godot 4.4.beta1
2. Create a new project
3. Create C# script and attach to a node in the main scene
4. Update the visual studio project to target .NET 9:
` <TargetFramework>net9.0</TargetFramework>`
4. Remote Deploy to Android with Gradle Build
### Minimal reproduction project (MRP)
N/A | bug,platform:android,topic:dotnet,regression,topic:export | low | Critical |
2,805,951,390 | pytorch | Incomplete check of LR as a tensor in Optimizer | ### 🐛 Describe the bug
A tutorial "[Running the compiled optimizer with an LR Scheduler](https://pytorch.org/tutorials/recipes/compiling_optimizer_lr_scheduler.html)" presents how LR as a tensor is used. According to this tutorial, we should use a 0-dim tensor for LR. However, Optimizer can accept a 1-dim tensor of size 1. When using a 1-dim tensor, we get a runtime error.
```
optimizer = torch.optim.SGD(model.parameters(), lr=torch.tensor([0.01]).cuda())
scheduler = torch.optim.lr_scheduler.LinearLR(optimizer, total_iters=5)
```
```
RuntimeError: fill_ only supports 0-dimension value tensor but got tensor with 1 dimensions.
```
Currently, Optimizer checks the tensor size:
https://github.com/pytorch/pytorch/blob/faa10faa2cad1cf6eef95a3e5bda255be6bc4c87/torch/optim/adam.py#L50-L56
I think that Optimizer should check the tensor dimension:
```
if lr.dim() != 0:
raise ValueError("Tensor lr must be 0-dimension")
```
The whole code I used:
```
import torch
print(torch.__version__)
# exit cleanly if we are on a device that doesn't support ``torch.compile``
if torch.cuda.get_device_capability() < (7, 0):
print("Exiting because torch.compile is not supported on this device.")
import sys
sys.exit(0)
# Setup random seed
torch.manual_seed(12345)
# Create simple model
model = torch.nn.Sequential(
*[torch.nn.Linear(1024, 1024, False, device="cuda") for _ in range(10)]
)
input = torch.rand(1024, device="cuda")
# run forward pass
output = model(input)
# run backward to populate the grads for our optimizer below
output.sum().backward()
optimizer = torch.optim.SGD(model.parameters(), lr=torch.tensor([0.01]).cuda())
scheduler = torch.optim.lr_scheduler.LinearLR(optimizer, total_iters=5)
def fn():
optimizer.step()
scheduler.step()
opt_fn = torch.compile(fn)
for i in range(5):
opt_fn()
print(i, optimizer.param_groups[0]['lr'].item())
```
### Error logs
```
---------------------------------------------------------------------------
RuntimeError Traceback (most recent call last)
<ipython-input-2-67dacbf46b2f> in <cell line: 26>()
24
25 optimizer = torch.optim.SGD(model.parameters(), lr=torch.tensor([0.01]).cuda())
---> 26 scheduler = torch.optim.lr_scheduler.LinearLR(optimizer, total_iters=5)
27
28 def fn():
/usr/local/lib/python3.10/dist-packages/torch/optim/lr_scheduler.py in __init__(self, optimizer, start_factor, end_factor, total_iters, last_epoch, verbose)
746 self.end_factor = end_factor
747 self.total_iters = total_iters
--> 748 super().__init__(optimizer, last_epoch, verbose)
749
750 def get_lr(self):
/usr/local/lib/python3.10/dist-packages/torch/optim/lr_scheduler.py in __init__(self, optimizer, last_epoch, verbose)
144 patch_track_step_called(self.optimizer)
145 self.verbose = _check_verbose_deprecated_warning(verbose)
--> 146 self._initial_step()
147
148 def _initial_step(self):
/usr/local/lib/python3.10/dist-packages/torch/optim/lr_scheduler.py in _initial_step(self)
149 """Initialize step counts and perform a step."""
150 self._step_count = 0
--> 151 self.step()
152
153 def state_dict(self):
/usr/local/lib/python3.10/dist-packages/torch/optim/lr_scheduler.py in step(self, epoch)
248 param_group, lr = data
249 if isinstance(param_group["lr"], Tensor):
--> 250 param_group["lr"].fill_(lr)
251 else:
252 param_group["lr"] = lr
RuntimeError: fill_ only supports 0-dimension value tensor but got tensor with 1 dimensions.
```
### Versions
PyTorch version: 2.5.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.3 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
Clang version: 14.0.0-1ubuntu1.1
CMake version: version 3.31.2
Libc version: glibc-2.35
Python version: 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0] (64-bit runtime)
Python platform: Linux-6.6.56+-x86_64-with-glibc2.35
Is CUDA available: True
CUDA runtime version: 12.2.140
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration:
GPU 0: Tesla T4
GPU 1: Tesla T4
Nvidia driver version: 560.35.03
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.9.6
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.9.6
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.9.6
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.9.6
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.9.6
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.9.6
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.9.6
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 46 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 4
On-line CPU(s) list: 0-3
Vendor ID: GenuineIntel
Model name: Intel(R) Xeon(R) CPU @ 2.00GHz
CPU family: 6
Model: 85
Thread(s) per core: 2
Core(s) per socket: 2
Socket(s): 1
Stepping: 3
BogoMIPS: 4000.30
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology nonstop_tsc cpuid tsc_known_freq pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch pti ssbd ibrs ibpb stibp fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm mpx avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves arat md_clear arch_capabilities
Hypervisor vendor: KVM
Virtualization type: full
L1d cache: 64 KiB (2 instances)
L1i cache: 64 KiB (2 instances)
L2 cache: 2 MiB (2 instances)
L3 cache: 38.5 MiB (1 instance)
NUMA node(s): 1
NUMA node0 CPU(s): 0-3
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Mitigation; PTE Inversion
Vulnerability Mds: Mitigation; Clear CPU buffers; SMT Host state unknown
Vulnerability Meltdown: Mitigation; PTI
Vulnerability Mmio stale data: Vulnerable: Clear CPU buffers attempted, no microcode; SMT Host state unknown
Vulnerability Reg file data sampling: Not affected
Vulnerability Retbleed: Mitigation; IBRS
Vulnerability Spec rstack overflow: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; IBRS; IBPB conditional; STIBP conditional; RSB filling; PBRSB-eIBRS Not affected; BHI SW loop, KVM SW loop
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Mitigation; Clear CPU buffers; SMT Host state unknown
Versions of relevant libraries:
[pip3] mypy-extensions==1.0.0
[pip3] numpy==1.26.4
[pip3] nvidia-cublas-cu12==12.6.4.1
[pip3] nvidia-cuda-cupti-cu12==12.6.80
[pip3] nvidia-cuda-runtime-cu12==12.6.77
[pip3] nvidia-cudnn-cu12==9.6.0.74
[pip3] nvidia-cufft-cu12==11.3.0.4
[pip3] nvidia-curand-cu12==10.3.7.77
[pip3] nvidia-cusolver-cu12==11.7.1.2
[pip3] nvidia-cusparse-cu12==12.5.4.2
[pip3] nvidia-nccl-cu12==2.23.4
[pip3] nvidia-nvjitlink-cu12==12.6.85
[pip3] nvtx==0.2.10
[pip3] onnx==1.17.0
[pip3] optree==0.13.1
[pip3] pynvjitlink-cu12==0.4.0
[pip3] pytorch-ignite==0.5.1
[pip3] pytorch-lightning==2.5.0.post0
[pip3] torch==2.5.1+cu121
[pip3] torchaudio==2.5.1+cu121
[pip3] torchinfo==1.8.0
[pip3] torchmetrics==1.6.1
[pip3] torchsummary==1.5.1
[pip3] torchtune==0.5.0
[pip3] torchvision==0.20.1+cu121
[conda] Could not collect
cc @vincentqb @jbschlosser @albanD @janeyx99 @crcrpar @chauhang @penguinwu | module: optimizer,triaged,actionable | low | Critical |
2,805,959,485 | vscode | Squash Merge will list a lot of commits msg that lead to crash in some cases | <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
- VS Code Version: 1.96.3
- OS Version: MacOS 14.5 (23F79)
Steps to Reproduce:
Assumes that have so many commits and need to merge into a branch with squash mode.
| info-needed,git | low | Critical |
2,805,981,004 | flutter | Impeller Rendering failed | ### Steps to reproduce
The following problems occur when using texture createSurfaceProducer on Huawei p30 devices in flutter-3.27.3 environment:
1-Impeller surfaceProducer is not enabled?. setSize(width, width * 9 / 16) Setting texture size has no effect, Bitmap crashes using textured surface generation

2-Turn on Impeller render blur and second creation crashes

### Expected results
...
### Actual results
...
### Code sample
<details open><summary>Code sample</summary>
```dart
private var width: Int = 0
if (Build.VERSION.SDK_INT >= Build.VERSION_CODES.R) {
val windowManager =
this.flutterPluginBinding!!.applicationContext.getSystemService(Context.WINDOW_SERVICE) as WindowManager
val windowMetrics = windowManager.currentWindowMetrics
val bounds = windowMetrics.bounds
width = bounds.width()
} else {
val metrics = DisplayMetrics()
(this.flutterPluginBinding!!.applicationContext.getSystemService(Context.WINDOW_SERVICE) as WindowManager).defaultDisplay.getMetrics(
metrics
)
width = metrics.widthPixels
}
val createSurfaceProducer = this.flutterPluginBinding?.textureRegistry?.createSurfaceProducer()
surfaceProducer = createSurfaceProducer
surfaceProducer?.setSize(width, width * 9 / 16)
```
</details>
### Screenshots or Video
The following problems occur when using texture createSurfaceProducer on Huawei p30 devices in flutter-3.27.3 environment:
1-Impeller surfaceProducer is not enabled?. setSize(width, width * 9 / 16) Setting texture size has no effect, Bitmap crashes using textured surface generation

2-Turn on Impeller render blur and second creation crashes

### Logs
<details open><summary>Logs</summary>
```console
D/permissions_handler(19155): No android specific permissions needed for: 6
D/permissions_handler(19155): No permissions found in manifest for: []9
I/flutter (19155): [IMPORTANT:flutter/shell/platform/android/platform_view_android.cc(308)] Flutter recommends migrating plugins that create and register surface textures to the new surface producer API. See https://docs.flutter.dev/release/breaking-changes/android-surface-plugins
V/version (19155): external version is 20240531_18:53:37_44aea9b4
I/AliFrameWork(19155): [0.9] [] :Ffmpeg version n4.3.1-8-gf381f263fe
E/AliFrameWork(19155): [0.9] [JNI] :0328 JNI_OnLoad
I/.saida.hengloo(19155): Thread[1,tid=19155,Native,Thread*=0x76e4c10800,peer=0x733ba540,"main"] recursive attempt to load library "/data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!/lib/arm64-v8a/libCicadaPlayer.so"
I/AliFrameWork(19155): [0.9] [] :setWorkPath :/data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/lib/arm64/
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :preferAudio 0
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :preferAudio 0
W/Gralloc3(19155): allocator 3.x is not supported
I/AliFrameWork(19155): [0.9] [] :AliyunCorePlayer jni_onVideoSizeChanged width = 2880 ,height = 1620
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :mPtsDiscontinueDelta = 20000000
I/AliFrameWork(19155): [0.9] [AbrBufferAlgoStrategy] :BA already change to bitrate:0
E/AliFrameWork(19155): [0.9] [SMPAVDeviceManager] :config decoder error ret= -513
E/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :setUpAudioDecoder error -514
E/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :setUpAVPath SetUpAudioPath failed,url is rtmp://v-jiangyin3-rtmp-play.videiot.cn:1937/live/101532044-1?eyJrZXkiOjk5OTksInNpZ24iOiJPLWt3ZzU4WHV1UENrMS1NeVVsV2xsWm9YMkhqeVhBV2dHZERuWmhGWjg1UDg2RFpUcFhDTXFwUFNGV3NKZTB5Q1EwWFZBUTdtaWlQcUI3eHNhTEI1bUZOTFVtUWtqNk1FdUkyMTFqbko3SXh5dXhaaFZmUHBYYTNmdXdmc1hQUjljRXRYRDZkR2NXMVpFWUhDZGtJUlhMSjJLNWp1SmVVYkFVeUxoWTRrdTgzZ1BmQjVzVkRpX0RzM3NsRWNnUzdIMm1HdnZYd05NUUs2eWg1YzRQdVl4bThZRjNUT3NaQ2R3RVhBOFVwM2JzIn0 video decoder open error
W/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :close audio stream
E/AliFrameWork(19155): [0.9] [GLRender_OESContext] :compileShader mVertShader failed. ret = -1
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :function name eglPresentationTimeANDROID
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :sym 0x7767517bac, critical true
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL client major 1 minor 4
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL extension EGL_KHR_get_all_proc_addresses EGL_ANDROID_presentation_time EGL_KHR_swap_buffers_with_damage EGL_ANDROID_get_native_client_buffer EGL_ANDROID_front_buffer_auto_refresh EGL_ANDROID_get_frame_timestamps EGL_EXT_surface_SMPTE2086_metadata EGL_EXT_surface_CTA861_3_metadata EGL_KHR_image EGL_KHR_image_base EGL_EXT_image_gl_colorspace EGL_KHR_gl_colorspace EGL_KHR_gl_texture_2D_image EGL_KHR_gl_texture_cubemap_image EGL_KHR_gl_renderbuffer_image EGL_KHR_fence_sync EGL_KHR_create_context EGL_KHR_config_attribs EGL_KHR_surfaceless_context EGL_EXT_create_context_robustness EGL_ANDROID_image_native_buffer EGL_KHR_wait_sync EGL_ANDROID_recordable EGL_KHR_partial_update EGL_EXT_pixel_format_float EGL_KHR_mutable_render_buffer EGL_EXT_protected_content EGL_IMG_context_priority EGL_KHR_no_config_context
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL attr version 2
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL config num 1
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL choose best config
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :eglMakeCurrent: display(0x1) surface(0x764407ff80) context(0x764407ff00)
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext CreateSurface mDisplay 0x1 mConfig 0x76775a8008 window 0x764a5431f0
D/mali_winsys(19155): EGLint new_window_surface(egl_winsys_display *, void *, EGLSurface, EGLConfig, egl_winsys_surface **, EGLBoolean) returns 0x3000
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext eglCreateWindowSurface surface : 0x76327cf480
W/VideoCapabilities(19155): Unrecognized profile/level 0/0 for video/mpeg2
W/VideoCapabilities(19155): Unrecognized profile/level 0/2 for video/mpeg2
W/VideoCapabilities(19155): Unrecognized profile/level 0/3 for video/mpeg2
I/VideoCapabilities(19155): Unsupported profile 5 for video/mpeg2
I/chatty (19155): uid=10879(com.saida.henglook) ApsaraPlayerSer identical 2 lines
I/VideoCapabilities(19155): Unsupported profile 5 for video/mpeg2
W/VideoCapabilities(19155): Unrecognized profile/level 1/32 for video/mp4v-es
W/VideoCapabilities(19155): Unrecognized profile/level 32768/2 for video/mp4v-es
W/VideoCapabilities(19155): Unrecognized profile/level 32768/64 for video/mp4v-es
I/OMXClient(19155): IOmx service obtained
I/ACodec (19155): In onAllocateComponent create compenent, codec name: OMX.hisi.video.decoder.avc
D/SurfaceUtils(19155): connecting to surface 0x763bbba010, reason connectToSurface
I/MediaCodec(19155): [OMX.hisi.video.decoder.avc] setting surface generation to 19614721
D/SurfaceUtils(19155): disconnecting from surface 0x763bbba010, reason connectToSurface(reconnect)
D/SurfaceUtils(19155): connecting to surface 0x763bbba010, reason connectToSurface(reconnect)
E/ACodec (19155): [OMX.hisi.video.decoder.avc] setPortMode on output to DynamicANWBuffer failed w/ err -2147483648
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104
I/ACodec (19155): onStart
I/HwExtendedUtils(19155): Set to window composer mode as 2
I/ACodec (19155): gralloc usage: 0(OMX) => 0x2900(ACodec)
D/SurfaceUtils(19155): disconnecting from surface 0x763bbba010, reason setNativeWindowSizeFormatAndUsage
D/SurfaceUtils(19155): connecting to surface 0x763bbba010, reason setNativeWindowSizeFormatAndUsage
D/SurfaceUtils(19155): set up nativeWindow 0x763bbba010 for 2880x1620, color 0x30d, rotation 0, usage 0x2900
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Allocating 6 buffers from a native window of size 7363584 on output port
I/AliFrameWork(19155): [0.9] [mediaCodecDecoder] :send Frame mFlushState = 2. pts 2444137047000
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Now handling output port settings change
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Output port now disabled.
I/HwExtendedUtils(19155): Set to window composer mode as 2
I/ACodec (19155): gralloc usage: 0(OMX) => 0x2900(ACodec)
D/SurfaceUtils(19155): disconnecting from surface 0x763bbba010, reason setNativeWindowSizeFormatAndUsage
D/SurfaceUtils(19155): connecting to surface 0x763bbba010, reason setNativeWindowSizeFormatAndUsage
D/SurfaceUtils(19155): set up nativeWindow 0x763bbba010 for 2880x1632, color 0x30d, rotation 0, usage 0x2900
W/ACodec (19155): [OMX.hisi.video.decoder.avc] setting nBufferCountActual to 11 failed: -1010
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Allocating 10 buffers from a native window of size 7050240 on output port
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Output port now reenabled.
D/ACodec (19155): sendVideoFpsDataToiAware time:2444137047000 fps:-1 msg:-1
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :Player NotifyFirstFrame
D/ACodec (19155): sendVideoFpsDataToiAware time:2444137087000 fps:25 msg:25
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
V/AudioManager(19155): querySoundEffectsEnabled...
E/flutter (19155): [ERROR:flutter/shell/platform/android/surface_texture_external_texture_vk_impeller.cc(122)] Break on 'ImpellerValidationBreak' to inspect point of failure: Invalid external texture.
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :Player ReadPacket Stop
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :not opened
I/AliFrameWork(19155): [0.9] [MeidaPlayerUtil] :KPI test finish: total fps:23.1
I/AliFrameWork(19155): [0.9] [GLRender] :drop a frame pts = 2444140407000
E/AliFrameWork(19155): [0.9] [GLRender] :~GLRender
W/BufferQueueProducer(19155): [SurfaceTexture-0-19155-0]:1390: disconnect: not connected (req=1)
W/libEGL (19155): EGLNativeWindowType 0x764568c010 disconnect failed
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext eglDestroySurface eglSurface 0x76327cf480
I/MediaCodec(19155): kWhatFlush processing is complete
I/ACodec (19155): [OMX.hisi.video.decoder.avc] ExecutingState flushing now (codec owns 2/5 input, 5/10 output).
I/MediaCodec(19155): Start Processing kWhatFlushCompleted
I/AliFrameWork(19155): [0.9] [mediaCodecDecoder] :clearCache. ret 0, flush state 1
W/ACodec (19155): forcing OMX state to Idle when received shutdown in ExecutingState
I/MediaCodec(19155): start process kWhatStopCompleted
D/SurfaceUtils(19155): disconnecting from surface 0x763bbba010, reason disconnectFromSurface
======== Exception caught by image resource service ================================================
The following HttpExceptionWithStatus was thrown resolving an image codec:
HttpException: Invalid statusCode: 400, uri = https://openapi.h.reservehemu.com/rest/service/camera/xxxxS_001c2718c086/thumbnail/current?size=1920x1080&token=b5c7d8f084984e0ab45613ff8261ac9e&channelId=0
When the exception was thrown, this was the stack:
Image provider: CachedNetworkImageProvider("https://openapi.h.reservehemu.com/rest/service/camera/xxxxS_001c2718c086/thumbnail/current?size=1920x1080&token=b5c7d8f084984e0ab45613ff8261ac9e&channelId=0", scale: 1.0)
Image key: CachedNetworkImageProvider("https://openapi.h.reservehemu.com/rest/service/camera/xxxxS_001c2718c086/thumbnail/current?size=1920x1080&token=b5c7d8f084984e0ab45613ff8261ac9e&channelId=0", scale: 1.0)
====================================================================================================
D/permissions_handler(19155): No android specific permissions needed for: 6
D/permissions_handler(19155): No permissions found in manifest for: []9
I/flutter (19155): [IMPORTANT:flutter/shell/platform/android/platform_view_android.cc(308)] Flutter recommends migrating plugins that create and register surface textures to the new surface producer API. See https://docs.flutter.dev/release/breaking-changes/android-surface-plugins
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :preferAudio 0
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :preferAudio 0
I/AliFrameWork(19155): [0.9] [] :AliyunCorePlayer jni_onVideoSizeChanged width = 2880 ,height = 1620
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :mPtsDiscontinueDelta = 20000000
E/AliFrameWork(19155): [0.9] [SMPAVDeviceManager] :config decoder error ret= -513
E/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :setUpAudioDecoder error -514
E/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :setUpAVPath SetUpAudioPath failed,url is rtmp://v-jiangyin3-rtmp-play.videiot.cn:1937/live/101532044-1?eyJrZXkiOjk5OTksInNpZ24iOiJPLWt3ZzU4WHV1UENrMS1NeVVsV2xsWm9YMkhqeVhBV2dHZERuWmhGWjg0SEtqSkFzbnBrWXNySUQxOThYcm5PcENTMnlxbllBd1NKcGN4c3liNjFzMmU3ZWFHNFl6SkhaZXB3aXl2S18tYk51RkZUajJzRWMzRGgtRkVGX3Z5RjRlQ2E2NlJvZFBqaVE5UEozcWdMM1p4LTc3bnYzTEJ0MzdqX1NDVDhfejAxMTJ3YWMxUFdaY3Zhc1p5R3pVdWE3MWJSMHNJSzRYbWhpVzREN3ptaUE5QU1xVGZnM3VJMHU2dW5uQTBHQVpRIn0 video decoder open error
W/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :close audio stream
I/AliFrameWork(19155): [0.9] [AbrBufferAlgoStrategy] :BA already change to bitrate:0
E/AliFrameWork(19155): [0.9] [GLRender_OESContext] :compileShader mVertShader failed. ret = -1
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :Egl .so already loaded
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL client major 1 minor 4
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL extension EGL_KHR_get_all_proc_addresses EGL_ANDROID_presentation_time EGL_KHR_swap_buffers_with_damage EGL_ANDROID_get_native_client_buffer EGL_ANDROID_front_buffer_auto_refresh EGL_ANDROID_get_frame_timestamps EGL_EXT_surface_SMPTE2086_metadata EGL_EXT_surface_CTA861_3_metadata EGL_KHR_image EGL_KHR_image_base EGL_EXT_image_gl_colorspace EGL_KHR_gl_colorspace EGL_KHR_gl_texture_2D_image EGL_KHR_gl_texture_cubemap_image EGL_KHR_gl_renderbuffer_image EGL_KHR_fence_sync EGL_KHR_create_context EGL_KHR_config_attribs EGL_KHR_surfaceless_context EGL_EXT_create_context_robustness EGL_ANDROID_image_native_buffer EGL_KHR_wait_sync EGL_ANDROID_recordable EGL_KHR_partial_update EGL_EXT_pixel_format_float EGL_KHR_mutable_render_buffer EGL_EXT_protected_content EGL_IMG_context_priority EGL_KHR_no_config_context
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL attr version 2
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL config num 1
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGL choose best config
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :eglMakeCurrent: display(0x1) surface(0x76327cfd80) context(0x76327cf900)
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext CreateSurface mDisplay 0x1 mConfig 0x76775a8008 window 0x764571ec28
D/mali_winsys(19155): EGLint new_window_surface(egl_winsys_display *, void *, EGLSurface, EGLConfig, egl_winsys_surface **, EGLBoolean) returns 0x3000
I/AliFrameWork(19155): [0.9] [GLRender egl_context] :EGLContext eglCreateWindowSurface surface : 0x76327cff00
I/OMXClient(19155): IOmx service obtained
I/ACodec (19155): In onAllocateComponent create compenent, codec name: OMX.hisi.video.decoder.avc
D/SurfaceUtils(19155): connecting to surface 0x763a818010, reason connectToSurface
I/MediaCodec(19155): [OMX.hisi.video.decoder.avc] setting surface generation to 19614722
D/SurfaceUtils(19155): disconnecting from surface 0x763a818010, reason connectToSurface(reconnect)
D/SurfaceUtils(19155): connecting to surface 0x763a818010, reason connectToSurface(reconnect)
E/ACodec (19155): [OMX.hisi.video.decoder.avc] setPortMode on output to DynamicANWBuffer failed w/ err -2147483648
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104
I/ACodec (19155): onStart
I/HwExtendedUtils(19155): Set to window composer mode as 2
I/ACodec (19155): gralloc usage: 0(OMX) => 0x2900(ACodec)
D/SurfaceUtils(19155): disconnecting from surface 0x763a818010, reason setNativeWindowSizeFormatAndUsage
D/SurfaceUtils(19155): connecting to surface 0x763a818010, reason setNativeWindowSizeFormatAndUsage
D/SurfaceUtils(19155): set up nativeWindow 0x763a818010 for 2880x1620, color 0x30d, rotation 0, usage 0x2900
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Allocating 6 buffers from a native window of size 7363584 on output port
I/AliFrameWork(19155): [0.9] [mediaCodecDecoder] :send Frame mFlushState = 2. pts 2444150007000
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] got color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) err=0(NO_ERROR)
I/ACodec (19155): [OMX.hisi.video.decoder.avc] using color aspects (R:2(Limited), P:1(BT709_5), M:1(BT709_5), T:3(SMPTE170M)) and dataspace 0x104
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Now handling output port settings change
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Output port now disabled.
I/HwExtendedUtils(19155): Set to window composer mode as 2
I/ACodec (19155): gralloc usage: 0(OMX) => 0x2900(ACodec)
D/SurfaceUtils(19155): disconnecting from surface 0x763a818010, reason setNativeWindowSizeFormatAndUsage
D/SurfaceUtils(19155): connecting to surface 0x763a818010, reason setNativeWindowSizeFormatAndUsage
D/SurfaceUtils(19155): set up nativeWindow 0x763a818010 for 2880x1632, color 0x30d, rotation 0, usage 0x2900
W/ACodec (19155): [OMX.hisi.video.decoder.avc] setting nBufferCountActual to 11 failed: -1010
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Allocating 10 buffers from a native window of size 7050240 on output port
I/ACodec (19155): [OMX.hisi.video.decoder.avc] Output port now reenabled.
D/ACodec (19155): sendVideoFpsDataToiAware time:2444150007000 fps:60 msg:60
I/AliFrameWork(19155): [0.9] [ApsaraPlayerService] :Player NotifyFirstFrame
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [GLRender] :drop a frame pts = 2444161687000
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
I/AliFrameWork(19155): [0.9] [avFormatDemuxer] :AV_PKT_DATA_NEW_EXTRADATA
W/System (19155): A resource failed to call end.
I/chatty (19155): uid=10879(com.saida.henglook) FinalizerDaemon identical 1 line
W/System (19155): A resource failed to call end.
F/libc (19155): Fatal signal 11 (SIGSEGV), code 1 (SEGV_MAPERR), fault addr 0x660 in tid 19172 (FinalizerDaemon), pid 19155 (.saida.henglook)
*** *** *** *** *** *** *** *** *** *** *** *** *** *** *** ***
Build fingerprint: 'HUAWEI/ELE-AL00/HWELE:10/HUAWEIELE-AL00/10.1.0.162C00:user/release-keys'
Revision: '0'
ABI: 'arm64'
SYSVMTYPE: Maple
APPVMTYPE: Art
Timestamp: 2025-01-23 13:42:49+0800
pid: 19155, tid: 19172, name: FinalizerDaemon >>> com.saida.henglook <<<
uid: 10879
signal 11 (SIGSEGV), code 1 (SEGV_MAPERR), fault addr 0x660
Cause: null pointer dereference
x0 0000000000000660 x1 00000076d8dbb5d4 x2 000000006ff4fb74 x3 00000076cd45b000
x4 00000076d8dbc660 x5 00000076e3b12ddf x6 647260646b647116 x7 7f7f7f7f7f7f7f7f
x8 5fae25f0c64e4ce6 x9 5fae25f0c64e4ce6 x10 0000000000430000 x11 00000076e4bfc000
x12 00000076e473e224 x13 00000076e473e26c x14 00000076e473e2cc x15 0000000000000000
x16 000000776863dab0 x17 0000007769029250 x18 000000767e8c8000 x19 0000000000000660
x20 0000000000000000 x21 00000076cd45b000 x22 00000076d8dbb810 x23 00000076e3b12ddf
x24 0000000000000004 x25 00000076d8dbd020 x26 00000076cd45b0b0 x27 0000000000000001
x28 0000000000000000 x29 00000076d8dbb580
sp 00000076d8dbb570 lr 00000077685edfc4 pc 0000007769029250
backtrace:
#00 pc 00000000000d0250 /apex/com.android.runtime/lib64/bionic/libc.so (pthread_mutex_lock) (BuildId: b91c775ccc9b0556e91bc575a2511cd0)
#01 pc 00000000000acfc0 /system/lib64/libgui.so (android::ConsumerBase::abandon()+24) (BuildId: 38bb0c19e01050d5b9cf5882b40316d7)
#02 pc 00000000001c4b30 /system/lib64/libandroid_runtime.so (android::SurfaceTexture_release(_JNIEnv*, _jobject*)+80) (BuildId: a92c351fcfda20f10ff25c04ed924f5f)
#03 pc 00000000002e649c /system/framework/arm64/boot-framework.oat (art_jni_trampoline+124) (BuildId: 6e4646237361a21448639f9cbe5b64c978005ca6)
#04 pc 0000000000147334 /apex/com.android.runtime/lib64/libart.so (art_quick_invoke_stub+548) (BuildId: 617b4481d2d525502ced0629cac33c57)
#05 pc 00000000001561b4 /apex/com.android.runtime/lib64/libart.so (art::ArtMethod::Invoke(art::Thread*, unsigned int*, unsigned int, art::JValue*, char const*)+252) (BuildId: 617b4481d2d525502ced0629cac33c57)
#06 pc 00000000002fd900 /apex/com.android.runtime/lib64/libart.so (art::interpreter::ArtInterpreterToCompiledCodeBridge(art::Thread*, art::ArtMethod*, art::ShadowFrame*, unsigned short, art::JValue*)+384) (BuildId: 617b4481d2d525502ced0629cac33c57)
#07 pc 00000000002f8bd0 /apex/com.android.runtime/lib64/libart.so (bool art::interpreter::DoCall<false, false>(art::ArtMethod*, art::Thread*, art::ShadowFrame&, art::Instruction const*, unsigned short, art::JValue*)+912) (BuildId: 617b4481d2d525502ced0629cac33c57)
#08 pc 00000000005cd8fc /apex/com.android.runtime/lib64/libart.so (MterpInvokeDirect+400) (BuildId: 617b4481d2d525502ced0629cac33c57)
#09 pc 0000000000141914 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_direct+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#10 pc 00000000003cc2b4 /system/framework/framework.jar (android.graphics.SurfaceTexture.release)
#11 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57)
#12 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#13 pc 00000000004eccba [anon:dalvik-classes19.dex extracted in memory from /data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!classes19.dex] (io.flutter.embedding.engine.renderer.SurfaceTextureWrapper.release+14)
#14 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57)
#15 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#16 pc 00000000004ebc7e [anon:dalvik-classes19.dex extracted in memory from /data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!classes19.dex] (io.flutter.embedding.engine.renderer.FlutterRenderer$SurfaceTextureRegistryEntry.release+78)
#17 pc 00000000005cd060 /apex/com.android.runtime/lib64/libart.so (MterpInvokeInterface+1752) (BuildId: 617b4481d2d525502ced0629cac33c57)
#18 pc 0000000000141a14 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_interface+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#19 pc 00000000004ecab8 [anon:dalvik-classes19.dex extracted in memory from /data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!classes19.dex] (io.flutter.embedding.engine.renderer.SurfaceTextureSurfaceProducer.release+4)
#20 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57)
#21 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#22 pc 00000000004eca58 [anon:dalvik-classes19.dex extracted in memory from /data/app/com.saida.henglook-_meJN1CWRaE1x8Uzt3FeoA==/base.apk!classes19.dex] (io.flutter.embedding.engine.renderer.SurfaceTextureSurfaceProducer.finalize+16)
#23 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57)
#24 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#25 pc 00000000001b45aa /apex/com.android.runtime/javalib/core-libart.jar (java.lang.Daemons$FinalizerDaemon.doFinalize+22)
#26 pc 00000000005cdbfc /apex/com.android.runtime/lib64/libart.so (MterpInvokeDirect+1168) (BuildId: 617b4481d2d525502ced0629cac33c57)
#27 pc 0000000000141914 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_direct+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#28 pc 00000000001b469c /apex/com.android.runtime/javalib/core-libart.jar (java.lang.Daemons$FinalizerDaemon.runInternal+164)
#29 pc 00000000005cb860 /apex/com.android.runtime/lib64/libart.so (MterpInvokeVirtual+1432) (BuildId: 617b4481d2d525502ced0629cac33c57)
#30 pc 0000000000141814 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_virtual+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#31 pc 00000000001b439e /apex/com.android.runtime/javalib/core-libart.jar (java.lang.Daemons$Daemon.run+50)
#32 pc 00000000005cd060 /apex/com.android.runtime/lib64/libart.so (MterpInvokeInterface+1752) (BuildId: 617b4481d2d525502ced0629cac33c57)
#33 pc 0000000000141a14 /apex/com.android.runtime/lib64/libart.so (mterp_op_invoke_interface+20) (BuildId: 617b4481d2d525502ced0629cac33c57)
#34 pc 00000000000eabdc /apex/com.android.runtime/javalib/core-oj.jar (java.lang.Thread.run+8)
#35 pc 00000000002ce22c /apex/com.android.runtime/lib64/libart.so (_ZN3art11interpreterL7ExecuteEPNS_6ThreadERKNS_20CodeItemDataAccessorERNS_11ShadowFrameENS_6JValueEbb.llvm.10887373532384510885+320) (BuildId: 617b4481d2d525502ced0629cac33c57)
#36 pc 00000000005bc090 /apex/com.android.runtime/lib64/libart.so (artQuickToInterpreterBridge+1012) (BuildId: 617b4481d2d525502ced0629cac33c57)
#37 pc 0000000000150468 /apex/com.android.runtime/lib64/libart.so (art_quick_to_interpreter_bridge+88) (BuildId: 617b4481d2d525502ced0629cac33c57)
#38 pc 0000000000147334 /apex/com.android.runtime/lib64/libart.so (art_quick_invoke_stub+548) (BuildId: 617b4481d2d525502ced0629cac33c57)
#39 pc 00000000001561b4 /apex/com.android.runtime/lib64/libart.so (art::ArtMethod::Invoke(art::Thread*, unsigned int*, unsigned int, art::JValue*, char const*)+252) (BuildId: 617b4481d2d525502ced0629cac33c57)
#40 pc 00000000004d8820 /apex/com.android.runtime/lib64/libart.so (art::(anonymous namespace)::InvokeWithArgArray(art::ScopedObjectAccessAlreadyRunnable const&, art::ArtMethod*, art::(anonymous namespace)::ArgArray*, art::JValue*, char const*)+104) (BuildId: 617b4481d2d525502ced0629cac33c57)
#41 pc 00000000004d98b4 /apex/com.android.runtime/lib64/libart.so (art::InvokeVirtualOrInterfaceWithJValues(art::ScopedObjectAccessAlreadyRunnable const&, _jobject*, _jmethodID*, jvalue const*)+416) (BuildId: 617b4481d2d525502ced0629cac33c57)
#42 pc 000000000051ca8c /apex/com.android.runtime/lib64/libart.so (art::Thread::CreateCallback(void*)+1232) (BuildId: 617b4481d2d525502ced0629cac33c57)
#43 pc 00000000000cf7c0 /apex/com.android.runtime/lib64/bionic/libc.so (__pthread_start(void*)+36) (BuildId: b91c775ccc9b0556e91bc575a2511cd0)
#44 pc 00000000000721a8 /apex/com.android.runtime/lib64/bionic/libc.so (__start_thread+64) (BuildId: b91c775ccc9b0556e91bc575a2511cd0)
Lost connection to device.
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[!] Flutter (Channel stable, 3.27.3, on Microsoft Windows [版本 10.0.22631.4751], locale zh-CN)
• Flutter version 3.27.3 on channel stable at D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter
! Warning: `flutter` on your path resolves to D:\Android\AndroidSdk\flutter_windows_3.24.3-stable\flutter\bin\flutter, which is not inside your current Flutter SDK checkout at D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter. Consider adding D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter\bin to the front of your path.
! Warning: `dart` on your path resolves to D:\Android\AndroidSdk\flutter_windows_3.24.3-stable\flutter\bin\dart, which is not inside your current Flutter SDK checkout at D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter. Consider adding D:\Android\AndroidSdk\flutter_windows_3.27.3-stable\flutter\bin to the front of your path.
• Upstream repository https://github.com/flutter/flutter.git
• Framework revision c519ee916e (35 hours ago), 2025-01-21 10:32:23 -0800
• Engine revision e672b006cb
• Dart version 3.6.1
• DevTools version 2.40.2
• If those were intentional, you can disregard the above warnings; however it is recommended to use "git" directly to perform update checks and upgrades.
[√] Windows Version (Installed version of Windows is version 10 or higher)
[!] Android toolchain - develop for Android devices (Android SDK version 34.0.0)
• Android SDK at D:\Android\AndroidSdk
X cmdline-tools component is missing
Run `path/to/sdkmanager --install "cmdline-tools;latest"`
See https://developer.android.com/studio/command-line for more details.
X Android license status unknown.
Run `flutter doctor --android-licenses` to accept the SDK licenses.
See https://flutter.dev/to/windows-android-setup for more details.
[√] Chrome - develop for the web
• Chrome at C:\Users\wjia3\AppData\Local\Google\Chrome\Application\chrome.exe
[√] Visual Studio - develop Windows apps (Visual Studio 生成工具 2019 16.11.5)
• Visual Studio at C:\Program Files (x86)\Microsoft Visual Studio\2019\BuildTools
• Visual Studio 生成工具 2019 version 16.11.31729.503
• Windows 10 SDK version 10.0.19041.0
[√] Android Studio (version 2024.1)
• Android Studio at D:\Program Files\Android\Android Studio
• Flutter plugin can be installed from:
https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
https://plugins.jetbrains.com/plugin/6351-dart
• Java version OpenJDK Runtime Environment (build 17.0.10+0--11609105)
[√] Android Studio (version 2024.2)
• Android Studio at D:\Program Files\Android\Android Studio1
• Flutter plugin can be installed from:
https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
https://plugins.jetbrains.com/plugin/6351-dart
• Java version OpenJDK Runtime Environment (build 17.0.10+0--11609105)
[√] IntelliJ IDEA Ultimate Edition (version 2022.1)
• IntelliJ at D:\Program Files\JetBrains\IntelliJ IDEA 2022.1.4
• Flutter plugin can be installed from:
https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
https://plugins.jetbrains.com/plugin/6351-dart
[√] Connected device (4 available)
• ELE AL00 (mobile) • Q5S5T19603021160 • android-arm64 • Android 10 (API 29)
• Windows (desktop) • windows • windows-x64 • Microsoft Windows [版本 10.0.22631.4751]
• Chrome (web) • chrome • web-javascript • Google Chrome 132.0.6834.83
• Edge (web) • edge • web-javascript • Microsoft Edge 132.0.2957.115
[!] Network resources
X A network error occurred while checking "https://maven.google.com/": 信号灯超时时间已到
X A network error occurred while checking "https://github.com/": 信号灯超时时间已到
! Doctor found issues in 3 categories.
Process finished with exit code 0
```
</details>
| waiting for customer response,in triage | low | Critical |
2,805,989,403 | go | x/tools/go/ssa: TestGenericFunctionSelector fails with EADDR (netbsd) | ```
#!watchflakes
default <- pkg == "golang.org/x/tools/go/ssa" && test == "TestGenericFunctionSelector"
```
Issue created automatically to collect these failures.
Example ([log](https://ci.chromium.org/b/8725122664788146929)):
=== RUN TestGenericFunctionSelector
builder_test.go:799: err: fork/exec /home/swarming/.swarming/w/ir/x/w/goroot/bin/go: bad address: stderr:
--- FAIL: TestGenericFunctionSelector (0.74s)
— [watchflakes](https://go.dev/wiki/Watchflakes)
| NeedsInvestigation,Tools | low | Critical |
2,805,997,681 | go | cmd/compile: worse performance when built with pgo | ### Go version
go version go1.23.5 linux/amd64
### Output of `go env` in your module/workspace:
```shell
GO111MODULE=''
GOARCH='amd64'
GOBIN=''
GOCACHE='/home/transaction/.cache/go-build'
GOENV='/home/transaction/.config/go/env'
GOEXE=''
GOEXPERIMENT=''
GOFLAGS=''
GOHOSTARCH='amd64'
GOHOSTOS='linux'
GOINSECURE=''
GOMODCACHE='/home/transaction/go/pkg/mod'
GONOPROXY=''
GONOSUMDB=''
GOOS='linux'
GOPATH='/home/transaction/go'
GOPRIVATE=''
GOPROXY='https://proxy.golang.org,direct'
GOROOT='/home/transaction/go/pkg/mod/golang.org/[email protected]'
GOSUMDB='sum.golang.org'
GOTMPDIR=''
GOTOOLCHAIN='auto'
GOTOOLDIR='/home/transaction/go/pkg/mod/golang.org/[email protected]/pkg/tool/linux_amd64'
GOVCS=''
GOVERSION='go1.23.5'
GODEBUG=''
GOTELEMETRY='local'
GOTELEMETRYDIR='/home/transaction/.config/go/telemetry'
GCCGO='gccgo'
GOAMD64='v1'
AR='ar'
CC='gcc'
CXX='g++'
CGO_ENABLED='1'
GOMOD='/home/transaction/tidb/go.mod'
GOWORK=''
CGO_CFLAGS='-O2 -g'
CGO_CPPFLAGS=''
CGO_CXXFLAGS='-O2 -g'
CGO_FFLAGS='-O2 -g'
CGO_LDFLAGS='-O2 -g'
PKG_CONFIG='pkg-config'
GOGCCFLAGS='-fPIC -m64 -pthread -Wl,--no-gc-sections -fmessage-length=0 -ffile-prefix-map=/tmp/go-build338251587=/tmp/go-build -gno-record-gcc-switches'
```
### What did you do?
Building with [PGO](https://go.dev/doc/pgo), but result in performance regressions.
The build flag is:
```
CGO_ENABLED=1 GO111MODULE=on go build -tags codes -pgo=default.pgo -ldflags '-X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v9.0.0-alpha-206-gaccc969cd5" -X "github.com/pingcap/tidb/pkg/util/versioninfo.TiDBBuildTS=2025-01-23 06:01:16" -X "github.com/pingcap/tidb/pkg/util/versioninfo.TiDBGitHash=accc969cd5e95a992368249fcf86f11887e7b11a" -X "github.com/pingcap/tidb/pkg/util/versioninfo.TiDBGitBranch=pgo-read-only-dev5" -X "github.com/pingcap/tidb/pkg/util/versioninfo.TiDBEdition=Community" ' -o bin/tidb-server ./cmd/tidb-server
```
Then start the TiDB and run sysbench test, following is the test result, the QPS of building with PGO is lower.
workload | thread | Version | QPS
-- | -- | -- | --
batch_get | 64 | without-pgo | 23165
batch_get | 64 | with-pgo | 21077
</byte-sheet-html-origin><!--EndFragment-->
</byte-sheet-html-origin><!--EndFragment-->
### What did you see happen?
The performance of build with PGO is worse because `newstack` appeared during execution.
[profile.zip](https://github.com/user-attachments/files/18515784/profile.zip)
CPU profile of building without PGO:
<img width="3006" alt="Image" src="https://github.com/user-attachments/assets/70a9f8f5-921f-4264-8ca6-1d777d9b20ae" />
CPU profile of building with PGO:
<img width="3003" alt="Image" src="https://github.com/user-attachments/assets/d401353e-db38-4c6e-ba2b-9cae189ddf55" />
### What did you expect to see?
no `newstack` during execution.
Is there any way to avoid `newstack` during execution, such as set stack size when creating goroutine? | NeedsInvestigation,compiler/runtime,BugReport | low | Critical |
2,806,003,966 | PowerToys | 在edge上无法反馈 | ### Provide a description of requested docs changes
在标题栏输入任意键会跳转至出现错误页面 | Issue-Docs,Needs-Triage | low | Minor |
2,806,004,012 | PowerToys | 在edge上无法反馈 | ### Provide a description of requested docs changes
在标题栏输入任意键会跳转至出现错误页面 | Issue-Docs,Needs-Triage | low | Minor |
2,806,008,166 | PowerToys | 鼠标荧光笔无法置于 always on top上 | ### Provide a description of requested docs changes
在置顶窗口后鼠标移至置顶的窗口会导致荧光笔无法使用,而在置顶窗口之外的区域依旧能够使用
| Issue-Docs,Needs-Triage | low | Minor |
2,806,059,279 | pytorch | Mark Dynamic does not work for nn module constructor inputs | for the following code if we compile it we get the following graph :
code:
```
import torch
import torch._dynamo.config
from torch.utils.checkpoint import _is_compiling
@torch.compile()
class Y(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, x):
return x.view(-1, self._N_input, 192)
@torch.compile(dynamic=True)
class M(torch.nn.Module):
def __init__(self, input):
super().__init__()
self._N_input = input.size()[0]
def forward(self, x, z):
# input is [B, n * d]
x = x*2
x = x*5
x = x *3
y = Y()
y._N_input = self._N_input
x = y.forward(x)# [B, n, d]
x = x*20
x = x*30
x = x*43
return x
x = torch.randn(5, 3210, 192) # Shape [B=4, n*d=12]
num_inputs = torch.randn(3210)
m = M(num_inputs)
y1, z1 = 3210, 192 # y1 * z1 == 12
output1 = m(x, z1)
print(output1)
```
graph: Note L_self_N_input is Sym(s1)
```
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] TRACED GRAPH
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] ===== __compiled_fn_1 =====
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] /home/lsakka/pytorch/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module):
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] def forward(self, s0: "Sym(s0)", s1: "Sym(s1)", L_x_: "f32[s0, s1, 192][192*s1, 192, 1]cpu", L_self_N_input: "Sym(s1)"):
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] l_x_ = L_x_
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] l_self_n_input = L_self_N_input
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:751 in forward, code: x = x*2
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = l_x_ * 2; l_x_ = None
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:752 in forward, code: x = x*5
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_1: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x * 5; x = None
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:753 in forward, code: x = x *3
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_2: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_1 * 3; x_1 = None
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:737 in __init__, code: super().__init__()
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] _log_api_usage_once = torch._C._log_api_usage_once('python.nn_module'); _log_api_usage_once = None
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:740 in forward, code: return x.view(-1, self._N_input, 192)
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_3: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_2.view(-1, l_self_n_input, 192); x_2 = l_self_n_input = None
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:759 in forward, code: x = x*20
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_4: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_3 * 20; x_3 = None
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:760 in forward, code: x = x*30
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_5: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_4 * 30; x_4 = None
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:761 in forward, code: x = x*43
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_6: "f32[s0, s1, 192][192*s1, 192, 1]cpu" = x_5 * 43; x_5 = None
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] return (x_6,)
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:39:29.878000 2775135 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
```
Now if we want to use mark_dynamic instead, I :
1) remove(dynamic=True)
2) call mark dynamic as the following:
```
mport torch
import torch._dynamo.config
from torch.utils.checkpoint import _is_compiling
@torch.compile()
class Y(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, x):
return x.view(-1, self._N_input, 192)
@torch.compile()
class M(torch.nn.Module):
def __init__(self, input):
super().__init__()
self._N_input = input.size()[0]
def forward(self, x, z):
# input is [B, n * d]
x = x*2
x = x*5
x = x *3
y = Y()
y._N_input = self._N_input
x = y.forward(x)# [B, n, d]
x = x*20
x = x*30
x = x*43
return x
x = torch.randn(5, 3210, 192) # Shape [B=4, n*d=12]
num_inputs = torch.randn(3210)
torch._dynamo.decorators.mark_dynamic(x, 1)
torch._dynamo.decorators.mark_dynamic(num_inputs, 0)
m = M(num_inputs)
y1, z1 = 3210, 192 # y1 * z1 == 12
output1 = m(x, z1)
print(output1)
```
but this generate this graph, looks like mark_dynamic did not work on num_inputs
raise ConstraintViolationError(
torch.fx.experimental.symbolic_shapes.ConstraintViolationError: Constraints violated (L['x'].size()[1])! For more information, run with TORCH_LOGS="+dynamic".
- Not all values of RelaxedUnspecConstraint(L['x'].size()[1]) are valid because L['x'].size()[1] was inferred to be a constant (3210).
```
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] TRACED GRAPH
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] ===== __compiled_fn_1 =====
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] /home/lsakka/pytorch/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module):
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] def forward(self, L_x_: "f32[5, 3210, 192][616320, 192, 1]cpu"):
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] l_x_ = L_x_
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:751 in forward, code: x = x*2
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x: "f32[5, 3210, 192][616320, 192, 1]cpu" = l_x_ * 2; l_x_ = None
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:752 in forward, code: x = x*5
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_1: "f32[5, 3210, 192][616320, 192, 1]cpu" = x * 5; x = None
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:753 in forward, code: x = x *3
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_2: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_1 * 3; x_1 = None
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:737 in __init__, code: super().__init__()
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] _log_api_usage_once = torch._C._log_api_usage_once('python.nn_module'); _log_api_usage_once = None
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:740 in forward, code: return x.view(-1, self._N_input, 192)
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_3: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_2.view(-1, 3210, 192); x_2 = None
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:759 in forward, code: x = x*20
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_4: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_3 * 20; x_3 = None
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:760 in forward, code: x = x*30
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_5: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_4 * 30; x_4 = None
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] # File: /home/lsakka/pytorch/example.py:761 in forward, code: x = x*43
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] x_6: "f32[5, 3210, 192][616320, 192, 1]cpu" = x_5 * 43; x_5 = None
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code] return (x_6,)
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
V0122 22:42:15.613000 2829692 torch/_dynamo/output_graph.py:1353] [0/0] [__graph_code]
```
cc @chauhang @penguinwu @ezyang @bobrenjc93 | triaged,module: dynamic shapes | low | Critical |
2,806,171,351 | react | Title Unnecessary updates to defaultValue and checked properties in <textarea> and <input> elements | ### Description
It looks like React is updating the defaultValue of <textarea> elements and the checked property of <input> elements even when their values haven’t changed. This leads to unnecessary DOM mutations, which can impact performance, especially when there are many components that re-render frequently.
### Steps to Reproduce
Render a controlled <textarea> or <input> with defaultValue or checked set.
Trigger a re-render by updating some other unrelated props.
Notice that React still updates the defaultValue or checked properties, even though the values haven't changed.
### Expected Behavior
React should only update the defaultValue or checked properties when the values actually change, avoiding unnecessary DOM updates.
### Actual Behavior
The properties are being updated unnecessarily, even if the values remain unchanged, leading to redundant DOM mutations.
### Suggested Solution
React should add a check to only update the defaultValue and checked properties if their values have changed. This would reduce unnecessary updates and improve overall performance.
### Environment
React version: [Insert React version]
Browser: [Insert browser and version]
Operating System: [Insert OS name and version]
### Key Words / Search Terms
defaultValue, checked, textarea, input, performance, DOM mutations, unnecessary updates
### Repository Version
[v19.0.0](https://github.com/facebook/react/releases/tag/v19.0.0)
| Status: Unconfirmed | medium | Major |
2,806,172,873 | react | Bug: | The custom Jest matchers introduced for async operations in the React Test Renderer (toFlushAll, toFlushThrough, toFlushAndThrow, toClearYields) throw unexpected errors in specific scenarios involving yielded values. These errors prevent the test suite from running as expected when handling complex async updates.
React version:
18.2.0
## Steps To Reproduce
1.Install React Test Renderer and add the custom Jest matchers.
2.Write a test using the toFlushAll matcher on a component rendering async updates.
3.Run the test suite and observe the unexpected behavior during flushing.
Link to code example:
link : https://codesandbox.io/p/sandbox/new
code:
import React from "react";
import TestRenderer, { act } from "react-test-renderer";
describe("React Test Renderer Async Matchers", () => {
it("throws an unexpected error with toFlushAll", () => {
const MyComponent = () => {
const [state, setState] = React.useState(0);
React.useEffect(() => {
setTimeout(() => setState(1), 1000);
}, []);
return <div>{state}</div>;
};
const renderer = TestRenderer.create(<MyComponent />);
expect(() => {
act(() => {
jest.runAllTimers();
renderer.unstable_flushAll(); // Fails here with an unexpected error
});
}).not.toThrow();
});
});
## The current behavior
When using toFlushAll, the test fails unexpectedly with a TypeError, preventing it from completing the async operations. The issue occurs when the matcher processes components with multiple nested yields.
## The expected behavior
The toFlushAll matcher should flush all pending async updates and yield the correct values without throwing errors. It should handle components with nested async updates seamlessly. | Status: Unconfirmed | medium | Critical |
2,806,200,871 | flutter | [native assets] [testing] Cleanup fake in tests | I stumbled upon some weird looking code in some of our testing fakes:
https://github.com/flutter/flutter/blob/97ca57cf0823a1948cbb6646a04054550842080d/packages/flutter_tools/test/general.shard/isolated/fake_native_assets_build_runner.dart#L25-L26
https://github.com/flutter/flutter/blob/97ca57cf0823a1948cbb6646a04054550842080d/packages/flutter_tools/test/general.shard/isolated/fake_native_assets_build_runner.dart#L37-L68
`onBuild` takes a `BuildConfig` (single hook invocation), but outputs a `BuildResult` which is the aggregated of all hook invocations. The implementation then does a loop over all packages and simply overwrites the `result` if there is more than one package with native assets. Unsurprisingly, we don't have any invocations with more than one package with native assets.
This should be rewritten in one of the two following ways:
* `final BuildOutput? Function(BuildConfig)? onBuild;` and loop over all packages and combine the results, or
* `final BuildResult? Function()? onBuild` and don't loop over the packages.
If this should be a shallow mock, it should just record what it is invoked with and not do anything fancy. So I'm leaning towards option 2.
cc @mkustermann you added the loop over the packages (presumably to get access to a `BuildConfig`), any opinions? | P3,c: tech-debt,team-tool,triaged-tool | low | Minor |
2,806,204,015 | kubernetes | code-generator v1.32.x defaulter-gen error | ### What happened?
defaulter-gen `zz_generated.defaults.go` Content error.
example case: https://github.com/dongjiang1989/customapis/tree/main/pkg/apis/custom/v1
```yaml
package v1
import (
corev1 "k8s.io/api/core/v1"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
)
// +genclient
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
// TestType is a top-level type. A client is created for it.
type TestType struct {
metav1.TypeMeta `json:",inline"`
// +optional
metav1.ObjectMeta `json:"metadata,omitempty"`
// +optional
Spec TestSpec `json:"spec,omitempty"`
}
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
// TestTypeList is a top-level list type. The client methods for lists are automatically created.
// You are not supposed to create a separated client for this one.
type TestTypeList struct {
metav1.TypeMeta `json:",inline"`
// +optional
metav1.ListMeta `json:"metadata,omitempty"`
Items []TestType `json:"items"`
}
type TestSpec struct {
ReplicaSpecs map[string]*ReplicaSpec `json:"replicaSpecs"` --------> this part
}
// +k8s:openapi-gen=true
// +k8s:deepcopy-gen=true
type ReplicaSpec struct {
// +optional
Replicas *int32 `json:"replicas,omitempty"`
Template corev1.PodTemplateSpec `json:"template,omitempty"`
}
````
output:
```yaml
package v1
import (
corev1 "k8s.io/api/core/v1" ---------> This imported package is not required
runtime "k8s.io/apimachinery/pkg/runtime"
)
// RegisterDefaults adds defaulters functions to the given scheme.
// Public to allow building arbitrary schemes.
// All generated defaulters are covering - they call all nested defaulters.
func RegisterDefaults(scheme *runtime.Scheme) error {
scheme.AddTypeDefaultingFunc(&TestType{}, func(obj interface{}) { SetObjectDefaults_TestType(obj.(*TestType)) })
scheme.AddTypeDefaultingFunc(&TestTypeList{}, func(obj interface{}) { SetObjectDefaults_TestTypeList(obj.(*TestTypeList)) })
return nil
}
func SetObjectDefaults_TestType(in *TestType) {
SetDefaults_TestType(in)
}
func SetObjectDefaults_TestTypeList(in *TestTypeList) {
for i := range in.Items {
a := &in.Items[i]
SetObjectDefaults_TestType(a)
}
}
```
### What did you expect to happen?
defaulter-gen `zz_generated.defaults.go` Content error.
### How can we reproduce it (as minimally and precisely as possible)?
```bash
dongjiang@MacBook Pro:customapis $ pwd
/Users/dongjiang/Documents/go/src/github.com/dongjiang1989/customapis
dongjiang@MacBook Pro:customapis $ ./hack/update-codegen.sh
/Users/dongjiang/Documents/go/pkg/mod/k8s.io/[email protected]
Generating deepcopy code for 1 targets
Generating defaulter code for 1 targets
Generating openapi code for 1 targets
Generating applyconfig code for 1 targets
Generating client code for 1 targets
Generating lister code for 1 targets
Generating informer code for 1 targets
```
### Anything else we need to know?
_No response_
### Kubernetes version
<details>
```console
$ kubectl version
# paste output here
```
</details>
code-generator v1.32.1 and v1.32.0
### Cloud provider
<details>
</details>
### OS version
<details>
```console
# On Linux:
$ cat /etc/os-release
# paste output here
$ uname -a
# paste output here
# On Windows:
C:\> wmic os get Caption, Version, BuildNumber, OSArchitecture
# paste output here
```
</details>
### Install tools
<details>
</details>
### Container runtime (CRI) and version (if applicable)
<details>
</details>
### Related plugins (CNI, CSI, ...) and versions (if applicable)
<details>
</details>
| kind/bug,needs-sig,needs-triage | low | Critical |
2,806,213,306 | opencv | “cv2.CAP_PROP_POS_FRAMES” jump is incorrect | ### System Information
Name: opencv-python Version: 4.8.0.76
System: Ubuntu 20.04
Python Version: 3.11.9
### Detailed description
I use cap.set(cv2.CAP_PROP_POS_FRAMES, start_frame) to jump to the 360th frame, with a frame rate of 60, which is 6 seconds, but I found that the frame after the jump is the video frame at the 12th second. I have only found a problem in one video,Other video frame rates are all 30 or 29,
and there is no problem in other videos. What could be the reason? Below is my code:
cap = cv2.VideoCapture(video_path)
if not cap.isOpened():
print(f"Error: Unable to open video file {video_path}")
return
# 获取视频的FPS和总帧数
fps = cap.get(cv2.CAP_PROP_FPS)
total_frames = int(cap.get(cv2.CAP_PROP_FRAME_COUNT))
# 计算从第几帧开始
start_frame = int(fps * start_time)
end_frame = int(fps * end_time)
# 跳转到起始帧
cap.set(cv2.CAP_PROP_POS_FRAMES, start_frame)
frame_interval = round(fps / fps_target) # 每秒需要提取的帧数(15帧每秒)
### Steps to reproduce
Video privacy concerns make it inconvenient to upload, sorry
### Issue submission checklist
- [x] I report the issue, it's not a question
- [x] I checked the problem with documentation, FAQ, open issues, forum.opencv.org, Stack Overflow, etc and have not found any solution
- [ ] I updated to the latest OpenCV version and the issue is still there
- [x] There is reproducer code and related data files (videos, images, onnx, etc) | bug,category: videoio,incomplete | low | Critical |
2,806,216,788 | storybook | [Storybook 9]: Remove deprecations | The following deprecations will be removed in Storybook 9:
| Feature | Description | File Path |
|---------|-------------|-----------|
| Addon Backgrounds | features.backgroundsStoryGlobals | code/addons/backgrounds/src/preview.ts |
| Addon Viewport | features.viewportStoryGlobals | code/addons/viewport/src/preview.ts |
| Addon Viewport | style function type | code/addons/viewport/src/types.ts |
| Addon_TestProviderType | title, description | code/core/src/types/modules/addons.ts |
| Addon_TypesEnum | experimental_SIDEBAR_BOTTOM, experimental_SIDEBAR_TOP | code/core/src/types/modules/addons.ts |
| API_Layout | isToolshown | code/core/src/types/modules/api.ts |
| API_Notification | DeprecatedIconType | code/core/src/types/modules/api.ts |
| API_Provider | serverChannel | code/core/src/types/modules/api.ts |
| CLI | renderer svelte template | code/core/src/cli/helpers.ts |
| Codemod | the entire package -> move to automigrations | code/lib/codemod |
| component | ButtonProps (isLink, primary, secondary, tertiary, gray, inForm, small, outline, containsIcon) | code/core/src/components/components/Button/Button.tsx |
| component | IconButtonSkeleton | code/core/src/components/components/bar/button.tsx |
| component | Icons | code/core/src/components/components/icon/icon.tsx |
| component | Symbols | code/core/src/components/components/icon/icon.tsx |
| CSFFile | _fileName, _makeTitle | code/core/src/csf-tools/CsfFile.ts |
| DocsOptions | autodocs | code/core/src/types/modules/core-common.ts |
| event | TESTING_MODULE_RUN_ALL_REQUEST | code/core/src/core-events/index.ts |
| Lib Test | fn | code/lib/test/src/spy.ts |
| Lib Test | Mock | code/lib/test/src/spy.ts |
| Manager API | DeprecatedState | code/core/src/manager-api/root.tsx |
| Manager API | getAddonState | code/core/src/manager-api/modules/addons.ts |
| Manager API | setAddonState | code/core/src/manager-api/modules/addons.ts |
| parameters.a11y.manual | | |
| Portable Stories | Legacy Portable Stories format | code/renderers/react/src/__test__/portable-stories-legacy.test.tsx |
| Portable Stories | Consolidate this renderContext with Context in SB 9.0 | code/core/src/preview-api/modules/store/csf/portable-stories.ts |
| Portable Stories | globals (replaced by initialGlobals) | code/core/src/preview-api/modules/store/csf/portable-stories.ts |
| Preview API | addons, mockChannel | code/core/src/preview-api/index.ts |
| Preview API | fromId | code/core/src/preview-api/modules/store/StoryStore.ts |
| Preview API | getSetStoriesPayload | code/core/src/preview-api/modules/store/StoryStore.ts |
| Preview API | getStoriesJsonData | code/core/src/preview-api/modules/store/StoryStore.ts |
| Preview API | globals (replaced by initialGlobals) | code/core/src/preview-api/modules/store/csf/normalizeProjectAnnotations.ts |
| Preview API | raw | code/core/src/preview-api/modules/store/StoryStore.ts |
| Preview API | serverChannel | code/core/src/preview-api/modules/preview-web/Preview.tsx |
| Preview API | storyStore | code/core/src/preview-api/modules/preview-web/Preview.tsx |
| project annotations | renderToDOM | code/core/src/types/modules/story.ts |
| type | Addon_SidebarBottomType | code/core/src/types/modules/addons.ts |
| type | Addon_SidebarTopType | code/core/src/types/modules/addons.ts |
| type | frameworkToRenderer | code/core/src/cli/helpers.ts |
| type | subcomponents | code/core/src/types/modules/addons.ts |
| type | SupportedRenderers | code/core/src/cli/project_types.ts | | bug,needs triage | low | Minor |
2,806,238,913 | pytorch | [inductor][torchbench] Unsupported operator issue when running the torch_multimodal_clip model with batch size 4. | ### 🐛 Describe the bug
I'm conducting an accuracy test for the `torch_multimodal_clip `model. I noticed that the graph breaks count is 0 when using a batch size of 1. However, when increasing the batch size to 4, the graph breaks count rises to 3. Upon printing the graph break details, I discovered the unsupported operator issue occurring with batch size 4.
```shell
python benchmarks/dynamo/torchbench.py --accuracy --amp -dcpu --inference -n50 --inductor --timeout 9000 --dynamic-shapes --dynamic-batch-only --freezing --only torch_multimodal_clip --print-graph-breaks --explain --batch-size 4
```
```
loading model: 0it [00:02, ?it/s]
cpu eval torch_multimodal_clip
WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu]
WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu]
WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu]
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] Graph break in user code at /workspace/pytorch/torch/nn/modules/activation.py:1313
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] Reason: Unsupported: unsupported operator: aten._native_multi_head_attention.default (see https://docs.google.com/document/d/1GgvOe7C8_NVOMLOCwDaYV1mXXyHMXY7ExoewHqooxrs/edit#heading=h.64r4npvq0w0 for how to fix)
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] User code traceback:
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/benchmarks/dynamo/common.py", line 2782, in run_n_iterations
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] self.model_iter_fn(mod, inputs, collect_outputs=False)
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/benchmarks/dynamo/torchbench.py", line 444, in forward_pass
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] return mod(*inputs)
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/opt/conda/lib/python3.10/site-packages/torchmultimodal/models/clip/model.py", line 71, in forward
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] embeddings_a = self.encoder_a(features_a)
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/opt/conda/lib/python3.10/site-packages/torchmultimodal/models/clip/image_encoder.py", line 108, in forward
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] x = self.encoder(x)
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/torch/nn/modules/transformer.py", line 517, in forward
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] output = mod(
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/torch/nn/modules/transformer.py", line 913, in forward
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] x = x + self._sa_block(
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/torch/nn/modules/transformer.py", line 934, in _sa_block
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] x = self.self_attn(
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] File "/workspace/pytorch/torch/nn/modules/activation.py", line 1313, in forward
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks] return torch._native_multi_head_attention(
V0123 16:00:42.112000 10589 torch/_dynamo/symbolic_convert.py:460] [0/0] [__graph_breaks]
V0123 16:00:42.211000 10589 torch/_dynamo/symbolic_convert.py:467] [1/0] [__graph_breaks] Graph break (details suppressed) in user code at /workspace/pytorch/torch/nn/modules/activation.py:1313
V0123 16:00:42.211000 10589 torch/_dynamo/symbolic_convert.py:467] [1/0] [__graph_breaks] Reason: Unsupported: unsupported operator: aten._native_multi_head_attention.default (see https://docs.google.com/document/d/1GgvOe7C8_NVOMLOCwDaYV1mXXyHMXY7ExoewHqooxrs/edit#heading=h.64r4npvq0w0 for how to fix)
V0123 16:00:42.466000 10589 torch/_dynamo/symbolic_convert.py:467] [2/0] [__graph_breaks] Graph break (details suppressed) in user code at /workspace/pytorch/torch/nn/modules/activation.py:1313
V0123 16:00:42.466000 10589 torch/_dynamo/symbolic_convert.py:467] [2/0] [__graph_breaks] Reason: Unsupported: unsupported operator: aten._native_multi_head_attention.default (see https://docs.google.com/document/d/1GgvOe7C8_NVOMLOCwDaYV1mXXyHMXY7ExoewHqooxrs/edit#heading=h.64r4npvq0w0 for how to fix)
V0123 16:00:42.569000 10589 torch/_dynamo/symbolic_convert.py:467] [3/0] [__graph_breaks] Graph break (details suppressed) in user code at /workspace/pytorch/torch/nn/modules/activation.py:1313
V0123 16:00:42.569000 10589 torch/_dynamo/symbolic_convert.py:467] [3/0] [__graph_breaks] Reason: Unsupported: unsupported operator: aten._native_multi_head_attention.default (see https://docs.google.com/document/d/1GgvOe7C8_NVOMLOCwDaYV1mXXyHMXY7ExoewHqooxrs/edit#heading=h.64r4npvq0w0 for how to fix)
pass
WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu]
Dynamo produced 6 graphs covering 209 ops with 3 graph breaks (1 unique)
```
For reference, below is the log of batch size 1
```shell
python benchmarks/dynamo/torchbench.py --accuracy --amp -dcpu --inference -n50 --inductor --timeout 9000 --dynamic-shapes --dynamic-batch-only --freezing --only torch_multimodal_clip --print-graph-breaks --explain --batch-size 1
```
```
loading model: 0it [00:02, ?it/s]
cpu eval torch_multimodal_clip
WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu]
WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu]
WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu]
pass
WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu]
Dynamo produced 1 graphs covering 744 ops with 0 graph breaks (0 unique)
```
### Versions
PyTorch version: 2.7.0a0+git9ae35b8
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.5 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
Clang version: Could not collect
CMake version: version 3.26.4
Libc version: glibc-2.35
Python version: 3.10.16 | packaged by conda-forge | (main, Dec 5 2024, 14:16:10) [GCC 13.3.0] (64-bit runtime)
Python platform: Linux-5.15.0-130-generic-x86_64-with-glibc2.35
Is CUDA available: False
CUDA runtime version: No CUDA
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 52 bits physical, 57 bits virtual
Byte Order: Little Endian
CPU(s): 384
On-line CPU(s) list: 0-383
Vendor ID: GenuineIntel
Model name: Intel(R) Xeon(R) 6972P
CPU family: 6
Model: 173
Thread(s) per core: 2
Core(s) per socket: 96
Socket(s): 2
Stepping: 1
CPU max MHz: 3900.0000
CPU min MHz: 800.0000
BogoMIPS: 4800.00
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 invpcid_single cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities
Virtualization: VT-x
L1d cache: 9 MiB (192 instances)
L1i cache: 12 MiB (192 instances)
L2 cache: 384 MiB (192 instances)
L3 cache: 960 MiB (2 instances)
NUMA node(s): 2
NUMA node0 CPU(s): 0-95,192-287
NUMA node1 CPU(s): 96-191,288-383
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Reg file data sampling: Not affected
Vulnerability Retbleed: Not affected
Vulnerability Spec rstack overflow: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS Not affected; BHI BHI_DIS_S
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] bert_pytorch==0.0.1a4
[pip3] functorch==1.14.0a0+b71aa0b
[pip3] mypy==1.14.1
[pip3] mypy-extensions==1.0.0
[pip3] numpy==1.26.4
[pip3] onnx==1.17.0
[pip3] torch==2.7.0a0+git9ae35b8
[pip3] torch_geometric==2.4.0
[pip3] torchaudio==2.6.0a0+b6d4675
[pip3] torchdata==0.7.0a0+11bb5b8
[pip3] torchmultimodal==0.1.0b0
[pip3] torchtext==0.16.0a0+b0ebddc
[pip3] torchvision==0.19.0a0+d23a6e1
[conda] bert-pytorch 0.0.1a4 dev_0 <develop>
[conda] functorch 1.14.0a0+b71aa0b pypi_0 pypi
[conda] mkl 2025.0.0 h901ac74_941 conda-forge
[conda] mkl-include 2025.0.0 hf2ce2f3_941 conda-forge
[conda] numpy 1.26.4 pypi_0 pypi
[conda] torch 2.7.0a0+git9ae35b8 dev_0 <develop>
[conda] torch-geometric 2.4.0 pypi_0 pypi
[conda] torchaudio 2.6.0a0+b6d4675 pypi_0 pypi
[conda] torchdata 0.7.0a0+11bb5b8 pypi_0 pypi
[conda] torchmultimodal 0.1.0b0 pypi_0 pypi
[conda] torchtext 0.16.0a0+b0ebddc pypi_0 pypi
[conda] torchvision 0.19.0a0+d23a6e1 pypi_0 pypi
cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki @chauhang @penguinwu @cpuhrsch @bhosmer @drisspg @soulitzer @davidberard98 @YuqingJ | module: nn,triaged,oncall: pt2,module: sdpa | low | Critical |
2,806,242,455 | vscode | Source control not working after adding turborepo to project | Does this issue occur when all extensions are disabled?: Yes
- VS Code Version:
- Version: 1.96.1 (Universal)
- Commit: 42b266171e51a016313f47d0c48aca9295b9cbb2
- Date: 2024-12-17T17:50:05.206Z (1 mo ago)
- Electron: 32.2.6
- ElectronBuildId: 10629634
- Chromium: 128.0.6613.186
- Node.js: 20.18.1
- V8: 12.8.374.38-electron.0
- OS: Darwin arm64 23.5.0
- OS and Version: MacOs sonoma 14.5
### Bug Summary and Steps to Reproduce
Bug Summary: After we added turborepo to our project, source control stopped working properly. After making a change to any file if I go to gitlens and select the file then I see "The editor could not be opened because the file was not found."
clicking on try again button I see : "Unable to read file 'git:/Users/nikunjgoyal/work/app/apps/web/Dockerfile?{"path":"/Users/nikunjgoyal/work/app/apps/web/Dockerfile","ref":""}' (Error: Unable to resolve nonexistent file 'git:/Users/nikunjgoyal/work/app/apps/web/Dockerfile?{"path":"/Users/nikunjgoyal/work/app/apps/web/Dockerfile","ref":""}')"

Steps to reproduce:
1. Not sure how this can be reproduced
Expected behavior: I should be able to compare differences in source control
| info-needed,git | low | Critical |
2,806,249,840 | vscode | Git error dialog swap button functions | <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
When an error occurs from the Git hooks, a dialog appears with buttons:
- Open Git Log
- Show Command Output
- Cancel
The problem is that the button click events are swapped. Cancel works correctly, but the other 2 buttons are wrong.
- Open Git Log -> Shows the command output
- Show Command Output -> Opens the Git Log
The solution is simple: Either swap the button title or the button click events. But I'm not sure what component is showing this dialog. Either VSCode or [Husky](https://github.com/typicode/husky)? Background: I'm working on an Angular project using the Husky node module for Git hooks. The dialog appears when a hook task failed (e.g. linter detected a problem).

Version: 1.96.4 (user setup)
Electron: 32.2.6
Chromium: 128.0.6613.186
Node.js: 20.18.1
V8: 12.8.374.38-electron.0
OS: Windows_NT x64 10.0.26100
| info-needed,git | low | Critical |
2,806,250,364 | storybook | [Storybook 9]: Drop support for tooling/integrations/frameworks | ### Describe the bug
The official support and maintenance of the following third-party toolings/integrations/frameworks will be removed in Storybook 9:
| Category | Tool | Versions |
|----------|------|----------|
| JavaScript Engine | Node.js | v18 |
| Package Manager | npm | v8, v9 |
| Package Manager | yarn | v3 |
| Package Manager | pnpm | v7, v8 |
| Renderer | React | v16, v17 (up for discussion) |
| Renderer | Angular | v15, v16, v17 |
| Renderer | Lit | v1, v2 (up for discussion) |
| Builders | Vite | v4 |
| Language | Typescript | < v4.9 (4.9 introduces satisfies support) |
| Renderer | Svelte | < v4 | | bug,needs triage | low | Critical |
2,806,254,072 | electron | Calls to window.open in offscreen documents do not work | ### Preflight Checklist
- [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project.
- [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [x] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success.
### Electron Version
34.0.1
### What operating system(s) are you using?
Other Linux
### Operating System Version
6.12.6-1
### What arch are you using?
x64
### Last Known Working Electron version
_No response_
### Expected Behavior
`window.open` calls in offscreen documents created with `chrome.offscreen.createDocument` routes the window open request to the webcontents window open handler set with `setWindowOpenHandler` or just opens the window in a new `BrowserWindow` and returns the `Window` object.
### Actual Behavior
`window.open` does nothing and returns null.
### Testcase Gist URL
https://github.com/BrettCleary/electron_bug_example
### Additional Information
Please see https://github.com/BrettCleary/electron_bug_example for a simple, reproducible example. I don't see an option to add an extension folder in electron fiddle and there is too much setup for a gist.
Also I know that full extension support is not a goal of core electron developers, but given that `chrome.offscreen.createDocument` works, I would expect `window.open` to work in the offscreen doc.
If you could even point me in the right direction to start troubleshooting, I could try to open a PR with a fix as well 🙏. I would love the opportunity to help with core electron development 🚀. | platform/linux,bug :beetle:,34-x-y | low | Critical |
2,806,255,097 | rust | Long type name emitted if the type name is part of a suggestion | ### Code
```Rust
use diesel::prelude::*; //diesel = { version = "2.2.6", default-features = false, features = ["postgres"] }
diesel::table! {
foo (id) {
id -> Int4,
name -> Text,
}
}
fn main() {
diesel::debug_query(
&foo::table
.filter(foo::id.eq(42))
.order(foo::name.desc())
.limit(10)
.offset(42),
);
}
```
### Current output
```Shell
error[E0282]: type annotations needed
--> src/main.rs:11:5
|
11 | diesel::debug_query(
| ^^^^^^^^^^^^^^^^^^^ cannot infer type of the type parameter `DB` declared on the function `debug_query`
|
help: consider specifying the generic arguments
|
11 | diesel::debug_query::<DB, SelectStatement<FromClause<table>, query_builder::select_clause::DefaultSelectClause<FromClause<table>>, query_builder::distinct_clause::NoDistinctClause, query_builder::where_clause::WhereClause<diesel::expression::grouped::Grouped<diesel::expression::operators::Eq<columns::id, diesel::expression::bound::Bound<Integer, i32>>>>, query_builder::order_clause::OrderClause<diesel::expression::operators::Desc<columns::name>>, LimitOffsetClause<LimitClause<diesel::expression::bound::Bound<BigInt, i64>>, OffsetClause<diesel::expression::bound::Bound<BigInt, i64>>>>>(
| ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
```
### Desired output
```Shell
error[E0282]: type annotations needed
--> src/main.rs:11:5
|
11 | diesel::debug_query(
| ^^^^^^^^^^^^^^^^^^^ cannot infer type of the type parameter `DB` declared on the function `debug_query`
|
help: consider specifying the generic arguments
|
11 | diesel::debug_query::<DB, _>(
| ++
```
### Rationale and extra context
Emitting the query type there is not helpful for the users as it is long, makes the error message harder to read and ultimately is not even required to fix the code snippet. Fixing the code snippet only requires to specify the `DB` generic type there. Ideally rustc would be able to suggest possible types for `DB` there, given the existing constraint on `debug_query`. In the example above that's even only satisfied by one type (`diesel::pg::Pg`).
### Other cases
If you change the exact query that you pass to `debug_query` the output changes. You can make the query more complex and therefore get a much more complex and longer query type in the compiler output.
### Rust Version
```Shell
rustc 1.86.0-nightly (649b995a9 2025-01-22)
binary: rustc
commit-hash: 649b995a9febd658b2570160703dff6fdc038ab2
commit-date: 2025-01-22
host: x86_64-unknown-linux-gnu
release: 1.86.0-nightly
LLVM version: 19.1.7
```
### Anything else?
cc @estebank as they expressed interested in such cases | A-diagnostics,T-compiler | low | Critical |
2,806,269,401 | flutter | [BUG] Error: PathNotFoundException during symbolic link resolution on Linux (/proc/self/exe not found) | ### Steps to reproduce
1. Run the Flutter application on a Linux device.
2. Observe the error during startup.
### Expected results
The application should start without errors, and the `path_provider_linux` package should successfully resolve symbolic links for the application support path.
### Actual results
`PathNotFoundException: Cannot resolve symbolic links, path = '/proc/self/exe' (OS Error: No such file or directory, errno = 2)`
### Code sample
<details open><summary>Code sample</summary>
```dart
Future<void> main() async {
WidgetsFlutterBinding.ensureInitialized();
await init(); // Injection container initialization
runApp(MyApp());
}
init(){
final prefsWithCache = await SharedPreferencesWithCache.create(
cacheOptions: const SharedPreferencesWithCacheOptions(),
);
/// Register AppPreferences and its implementation
di.registerLazySingleton<AppPreferences>(
() => AppPreferencesImpl(
prefsWithCache,
),
);
}
class AppPreferencesImpl implements AppPreferences {
late SharedPreferencesWithCache _sharedPreferences;
set sharedPreferences(SharedPreferencesWithCache value) {
_sharedPreferences = value;
}
AppPreferencesImpl(SharedPreferencesWithCache sharedPrefs) {
_sharedPreferences = sharedPrefs;
initSharedPrefs();
}
initSharedPrefs() async {
_sharedPreferences = await SharedPreferencesWithCache.create(
cacheOptions: const SharedPreferencesWithCacheOptions(),
);
}
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>

</details>
### Logs
<details open><summary>Logs</summary>
```console
(innogec_mobile:161765): GLib-GObject-WARNING **: 09:28:46.135: g_object_weak_unref: couldn't find weak ref 0x75a497ff0b50(0x29c234d8)
Restarted application in 684ms.
[ERROR:flutter/runtime/dart_vm_initializer.cc(40)] Unhandled Exception: PathNotFoundException: Cannot resolve symbolic links, path = '/proc/self/exe' (OS Error: No such file or directory, errno = 2)
#0 _checkForErrorResponse (dart:io/common.dart:56:9)
common.dart:56
#1 FileSystemEntity.resolveSymbolicLinks.<anonymous closure> (dart:io/file_system_entity.dart:318:7)
file_system_entity.dart:318
<asynchronous suspension>
#2 PathProviderLinux._getExecutableName (package:path_provider_linux/src/path_provider_linux.dart:92:9)
path_provider_linux.dart:92
<asynchronous suspension>
#3 PathProviderLinux._getId (package:path_provider_linux/src/path_provider_linux.dart:100:30)
path_provider_linux.dart:100
<asynchronous suspension>
#4 PathProviderLinux.getApplicationSupportPath (package:path_provider_linux/src/path_provider_linux.dart:51:48)
path_provider_linux.dart:51
<asynchronous suspension>
#5 _getLocalDataFile (package:shared_preferences_linux/shared_preferences_linux.dart:339:29)
shared_preferences_linux.dart:339
<asynchronous suspension>
#6 _reload (package:shared_preferences_linux/shared_preferences_linux.dart:354:31)
shared_preferences_linux.dart:354
<asynchronous suspension>
#7 SharedPreferencesAsyncLinux._readPreferences (package:shared_preferences_linux/shared_preferences_linux.dart:323:28)
shared_preferences_linux.dart:323
<asynchronous suspension>
#8 SharedPreferencesAsyncLinux._readAll (package:shared_preferences_linux/shared_preferences_linux.dart:300:34)
shared_preferences_linux.dart:300
<asynchronous suspension>
#9 SharedPreferencesWithCache.reloadCache (package:shared_preferences/src/shared_preferences_async.dart:225:9)
shared_preferences_async.dart:225
<asynchronous suspension>
#10 SharedPreferencesWithCache.create (package:shared_preferences/src/shared_preferences_async.dart:200:5)
shared_preferences_async.dart:200
<asynchronous suspension>
#11 init (package:innogec_mobile/injection_container.dart:58:26)
injection_container.dart:58
<asynchronous suspension>
#12 mainCommon (package:innogec_mobile/main_common.dart:16:3)
main_common.dart:16
<asynchronous suspension>
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[✓] Android toolchain - develop for Android devices (Android SDK version 35.0.0)
• Android SDK at /home/user/Android/Sdk
• Platform android-35, build-tools 35.0.0
• Java binary at: /snap/android-studio/161/jbr/bin/java
• Java version OpenJDK Runtime Environment (build
17.0.10+0-17.0.10b1087.21-11609105)
• All Android licenses accepted.
[✓] Chrome - develop for the web
• Chrome at google-chrome
[✓] Linux toolchain - develop for Linux desktop
• clang version 10.0.0-4ubuntu1
• cmake version 3.16.3
• ninja version 1.10.0
• pkg-config version 0.29.1
[✓] Android Studio (version 2023.3)
• Android Studio at /snap/android-studio/157
• Flutter plugin version 79.0.2
• Dart plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/6351-dart
• Java version OpenJDK Runtime Environment (build
17.0.10+0-17.0.10b1087.21-11572160)
[✓] Android Studio (version 2024.1)
• Android Studio at /snap/android-studio/161
• Flutter plugin version 83.0.2
• Dart plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/6351-dart
• Java version OpenJDK Runtime Environment (build
17.0.10+0-17.0.10b1087.21-11609105)
[✓] IntelliJ IDEA Ultimate Edition (version 2024.3)
• IntelliJ at /snap/intellij-idea-ultimate/561
• Flutter plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/6351-dart
[✓] VS Code (version 1.96.4)
• VS Code at /snap/code/current/usr/share/code
• Flutter extension version 3.102.0
[✓] Connected device (2 available)
• Linux (desktop) • linux • linux-x64 • Ubuntu 22.04.5 LTS
6.8.0-51-generic
• Chrome (web) • chrome • web-javascript • Google Chrome 132.0.6834.83
[✓] Network resources
• All expected network resources are available.
• No issues found!
```
</details>
| waiting for customer response,in triage | low | Critical |
2,806,272,957 | vscode | Crashing when clicking on File menu with native title bar |
Type: <b>Bug</b>
Steps to Reproduce:
1. Arch linux with wayland (hyprland compositor)
2. Hover/click the file menu
2. All windows crash
This looks very similar to the issue discussed at https://github.com/microsoft/vscode/issues/230604
But at least for me, I'm on latest versions and made sure that the recommended workaroudn of removing meta/super key from the keybind is applied, still - vscode crashes for me.
Still happens if I launch with `code --disable-extensions`
If I downgrade to 1.94.0 it no longer crashes but I cannot use extensions such as GH copilot.
VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z)
OS version: Linux x64 6.12.10-arch1-1
Modes:
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) Ultra 7 165H (22 x 3804)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off<br>webnn: disabled_off|
|Load (avg)|4, 2, 2|
|Memory (System)|62.24GB (51.33GB free)|
|Process Argv|-enable-features=UseOzonePlatform,WaylandWindowDecorations --ozone-platform=wayland --crash-reporter-id 83481648-87d4-4456-9caf-70c8cdfed42f|
|Screen Reader|no|
|VM|0%|
|DESKTOP_SESSION|undefined|
|XDG_CURRENT_DESKTOP|Hyprland|
|XDG_SESSION_DESKTOP|Hyprland|
|XDG_SESSION_TYPE|wayland|
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368:30146709
vspor879:30202332
vspor708:30202333
vspor363:30204092
vswsl492cf:30256860
vscod805:30301674
binariesv615:30325510
vsaa593:30376534
py29gd2263:31024239
vscaat:30438848
c4g48928:30535728
azure-dev_surveyone:30548225
a9j8j154:30646983
962ge761:30959799
pythonnoceb:30805159
pythonmypyd1:30879173
2e7ec940:31000449
pythontbext0:30879054
cppperfnew:31000557
dsvsc020:30976470
pythonait:31006305
dsvsc021:30996838
dvdeprecation:31068756
dwnewjupytercf:31046870
nativerepl2:31139839
pythonrstrctxt:31112756
nativeloc2:31192216
cf971741:31144450
iacca1:31171482
notype1cf:31157160
5fd0e150:31155592
dwcopilot:31170013
stablechunks:31184530
6074i472:31201624
dwoutputs:31217127
hdaa2157:31222309
copilot_t_ci:31222730
```
</details>
<!-- generated by issue reporter --> | bug,upstream,freeze-slow-crash-leak,upstream-issue-linked,chromium,mitigated | low | Critical |
2,806,284,064 | transformers | resume_from_checkpoint failed when using PEFT LORA | ### System Info
- `transformers` version: 4.43.1
- Platform: Linux-6.5.0-45-generic-x86_64-with-glibc2.35
- Python version: 3.10.14
- Huggingface_hub version: 0.24.7
- Safetensors version: 0.4.5
- Accelerate version: 1.1.0
- Accelerate config: - compute_environment: LOCAL_MACHINE
- distributed_type: MULTI_GPU
- mixed_precision: no
- use_cpu: False
- debug: False
- num_processes: 2
- machine_rank: 0
- num_machines: 1
- gpu_ids: all
- rdzv_backend: static
- same_network: True
- main_training_function: main
- enable_cpu_affinity: False
- downcast_bf16: no
- tpu_use_cluster: False
- tpu_use_sudo: False
- tpu_env: []
- PyTorch version (GPU?): 2.4.1+cu124 (True)
- Tensorflow version (GPU?): not installed (NA)
- Flax version (CPU?/GPU?/TPU?): not installed (NA)
- Jax version: not installed
- JaxLib version: not installed
- Using distributed or parallel set-up in script?: <fill in>
- Using GPU in script?: <fill in>
- GPU type: NVIDIA H100 NVL
### Who can help?
@muellerzr @SunMarc
### Information
- [ ] The official example scripts
- [x] My own modified scripts
### Tasks
- [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...)
- [x] My own task or dataset (give details below)
### Reproduction
I am training my own model (a vision-language model) with LoRA using the Trainer. However, when I try to use resume_from_checkpoint, I encounter the following error:
```
Traceback (most recent call last):
File "/dataT0/Free/hengyue/dataF0_save/new/LMM/training_Trainer.py", line 137, in <module>
trainer.train(resume_from_checkpoint=True)
File "/dataF0/Free/hengyue/miniconda3/envs/doraemon24/lib/python3.10/site-packages/transformers/trainer.py", line 2143, in train
self._load_from_checkpoint(resume_from_checkpoint)
File "/dataF0/Free/hengyue/miniconda3/envs/doraemon24/lib/python3.10/site-packages/transformers/trainer.py", line 2851, in _load_from_checkpoint
if len(active_adapters) > 1:
TypeError: object of type 'method' has no len()
```
My Code Shows below:
```
lora_config = LoraConfig(
task_type=TaskType.CAUSAL_LM,
inference_mode=False,
r=8,
lora_alpha=32,
lora_dropout=0.05,
target_modules=["q_proj", "k_proj", "v_proj", "o_proj"],
)
model = MyModel()
model = get_peft_model(model, lora_config)
model.print_trainable_parameters()
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
train_dataset = Mydataset()
val_dataset = Valdataset()
training_args = TrainingArguments(
output_dir="./results",
evaluation_strategy="epoch",
save_strategy="epoch",
learning_rate=1e-4,
per_device_train_batch_size=2,
per_device_eval_batch_size=1,
num_train_epochs=11,
weight_decay=0.01,
logging_dir="./logs",
logging_steps=10,
save_total_limit=2,
report_to="wandb",
dataloader_num_workers = 4,
)
trainer = Trainer(
model=model,
args=training_args,
train_dataset=train_dataset,
eval_dataset=val_dataset,
data_collator=custom_collate_fn,
)
trainer.train(resume_from_checkpoint=True)
trainer.save_state()
trainer.save_model("./test_model")
```
### Expected behavior
Expected the model should continue training from the checkpoint. | bug | low | Critical |
2,806,293,420 | react | [DevTools Bug]: Support Iterables in React DevTools | ### Website or app
React DevTools
### Repro steps
1.Open a React app using React 19.0.0.
2.Pass an Iterable<T> to a component's props or state.
3.Inspect the component using React DevTools.
4.Observe the display or handling of the Iterable.
### How often does this bug happen?
Every time
### DevTools package (automated)
_No response_
### DevTools version (automated)
_No response_
### Error message (automated)
_No response_
### Error call stack (automated)
```text
```
### Error component stack (automated)
```text
```
### GitHub query string (automated)
```text
``` | Type: Bug,Status: Unconfirmed,Component: Developer Tools | medium | Critical |
2,806,304,172 | terminal | The lowercase I glyph (U+0069) in JuliaMono renders incorrectly | ### Windows Terminal version
1.22.241118002-preview, also happens on 1.21.3231.0
### Windows build number
10.0.26100.2894
### Other Software
_No response_
### Steps to reproduce
1. Download the JuliaMono font if you haven't yet.
2. Set the terminal font to JuliaMono.
3. Open a terminal and type in "i". The glyph renders incorrectly for me.
### Expected Behavior
I was expecting this:

Note the serifs in the lowercase I.
### Actual Behavior
I got this behavior:

Some notes:
- Other variations of the lowercase I (î, ï, ı etc.) render with serifs.
- Uppercase I renders with serifs, as seen in the picture. | Issue-Bug,Needs-Triage | low | Major |
2,806,313,473 | react | Clarify Behavior of useSyncExternalStore When Using Server-Side Rendering (SSR) | # Title:
Clarify Behavior of useSyncExternalStore When Using Server-Side Rendering (SSR)
## Description:
After reviewing PR #29760, which addresses the optimization of useSyncExternalStore in React, I noticed that the behavior of this hook when used in server-side rendering (SSR) scenarios isn't clearly documented. Specifically, the following points need clarification: How should developers manage subscriptions during SSR? What best practices should be followed to avoid potential inconsistencies when hydrating the client-side store?
<!--
Please provide a clear and concise description of what the bug is. Include
screenshots if needed. Please test using the latest version of the relevant
React packages to make sure your issue has not already been fixed.
-->
React version: 19
## Steps To Reproduce
1. Set up the environment: Include details about the environment where the issue occurs (e.g., React version, Node.js version, browser, etc.).
2. Create a minimal project: Describe how to create a minimal reproducible example that isolates the bug (e.g., a simple React app using create-react-app).
3. Write relevant code: Provide the key parts of the code where the issue happens (e.g., using a specific React feature, hook, or component).
4. Run the project: Explain how to execute the project to trigger the bug.
5. Observe the behavior: Describe what happens when the steps are followed (the actual behavior), compared to what you expected to happen.
<!--
Your bug will get fixed much faster if we can run your code and it doesn't
have dependencies other than React. Issues without reproduction steps or
code examples may be immediately closed as not actionable.
-->
Link to code example:
<!--
Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a
repository on GitHub, or provide a minimal code example that reproduces the
problem. You may provide a screenshot of the application if you think it is
relevant to your bug report. Here are some tips for providing a minimal
example: https://stackoverflow.com/help/mcve.
-->
Please use one of the following techniques to produce a simple, replicable sample of the bug:
Give a link to a CodeSandbox project that illustrates the problem.
GitHub Repository: Provide a link to a small project on GitHub that replicates the issue.
Inline Code: Use appropriate formatting to include the code right in the bug report if it is brief and simple.
## The current behavior
There isn't much advice on SSR-specific circumstances in the present material.
## The expected behavior
The documentation should explicitly detail how useSyncExternalStore interacts with SSR.
Provide examples of correct usage in SSR and hydration contexts.
## References
- Related Pull Request: [29760](https://github.com/facebook/react/pull/29760) | Status: Unconfirmed | medium | Critical |
2,806,323,057 | godot | There is no order in calling method _input_event() of Area2D. Expectation is order by node position in sence tree | ### Tested versions
4.3-stable
### System information
MacOS 15
### Issue description
See the picture below, i have created a minimal reproduction project following.
This issue make it impossible to control the propagation order of input event.

### Steps to reproduce
NA
### Minimal reproduction project (MRP)
[buttest.zip](https://github.com/user-attachments/files/18517822/buttest.zip) | discussion,documentation,topic:physics | low | Minor |
2,806,325,434 | bitcoin | test: intermittent issue in p2p_1p1c_network.py | Looks like this still happens: https://github.com/bitcoin/bitcoin/pull/31713#issuecomment-2609212276
So this is likely a non-timeout issue. Note that the test has immediate tx-relay (`self.noban_tx_relay`) enabled, so trickle delay should not be an issue.
_Originally posted by @maflcko in https://github.com/bitcoin/bitcoin/issues/31701#issuecomment-2609226329_
| CI failed | low | Major |
2,806,330,642 | vscode | VSCode will rollback file content |
Type: <b>Bug</b>
I am using vs2022 and vscode for same folder, but vscode will rollback other editor's change.
First use other editor change the file, save, then use vs code open it, the file will be rollback, I tested , the reason is vs code will use cache "C:\Users\qwe\AppData\Roaming\Code", if delete this folder, it will fix, but vs code do not provide a option to disable the "cache",
Please fix the bug, or there is another way to fix this please tell me , thank you!
VS Code version: Code 1.96.4 (cd4ee3b1c348a13bafd8f9ad8060705f6d4b9cba, 2025-01-16T00:16:19.038Z)
OS version: Windows_NT x64 10.0.19045
Modes:
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) N100 (4 x 806)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off|
|Load (avg)|undefined|
|Memory (System)|15.68GB (3.70GB free)|
|Process Argv|--crash-reporter-id 55627040-681b-4624-ae9b-2941d63e8d17|
|Screen Reader|no|
|VM|40%|
</details><details><summary>Extensions (9)</summary>
Extension|Author (truncated)|Version
---|---|---
copilot|Git|1.258.0
copilot-chat|Git|0.23.2
better-cpp-syntax|jef|1.27.1
marscode-extension|Mar|1.1.58
cmake-tools|ms-|1.19.52
cpptools|ms-|1.22.11
cpptools-extension-pack|ms-|1.3.0
BuildOutputColorizer|Ste|0.1.8
cmake|twx|0.0.17
(1 theme extensions excluded)
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vspor879:30202332
vspor708:30202333
vspor363:30204092
vswsl492cf:30256860
vscod805:30301674
binariesv615:30325510
vsaa593cf:30376535
py29gd2263:31024239
c4g48928:30535728
azure-dev_surveyone:30548225
962ge761:30959799
pythonnoceb:30805159
pythonmypyd1:30879173
h48ei257:31000450
pythontbext0:30879054
cppperfnew:31000557
dsvsc020:30976470
pythonait:31006305
dsvsc021:30996838
dvdeprecation:31068756
dwnewjupytercf:31046870
2f103344:31071589
nativerepl1:31139838
pythonrstrctxt:31112756
nativeloc2:31192216
cf971741:31144450
iacca1:31171482
notype1:31157159
5fd0e150:31155592
dwcopilot:31170013
stablechunks:31184530
6074i472:31201624
dwoutputs:31217127
9064b325:31222308
copilot_t_ci:31222730
```
</details>
<!-- generated by issue reporter --> | workbench-copilot | low | Critical |
2,806,376,723 | rust | Cleanup "no panic support" tests | There are a bunch of tests that say `ignore-wasm` or `ignore-emscripten` or `ignore-$platform` with a comment "no panic support" or "no panic". It would be good to audit them and see if one or more of the following capability-based conditional test execution directives can be used instead:
- `//@ needs-unwind`
- `//@ needs-threads`
- `//@ needs-subprocess`[^1]
If those are not sufficient, it would be good to improve the ignore reason anyway.
[^1]: I'm still working on implementing this directive over at #128295. | C-cleanup,A-testsuite,T-compiler,E-medium,A-panic | low | Minor |
2,806,379,010 | flutter | FFI: SIGSEGV crash in release mode when handling image data between C++ and Dart, woks in debug mode | ### Steps to reproduce
Clone repository: https://github.com/rmatif/Local-Diffusion
Build and run in debug mode - works correctly
Build release APK: flutter build apk --release --target-platform android-arm64
Install and run release APK
Initialize model and generate an image
Observe crash after image generation completes
Already attempted fixes:
- Added ProGuard rules to preserve FFI classes
<details open><summary>proguard-rules.pro</summary>
```pro
-keep class * extends ffi.Struct {
*;
}
-keep class * implements ffi.Finalizable {
*;
}
-keep class ffi_bindings.FFIBindings { # Keep the FFIBindings class itself
*; # Keep all members
}
-keep class sd_image.SDImage { # Keep the SDImage class
*; # Keep all members
}
-keep class com.example.local_diffusion.** { # Keep your main app package. Be more specific if possible.
*;
}
# If you have specific callback functions that are only called from native code:
-keepclassmembers class ffi_bindings.FFIBindings {
public static void _staticLogCallback(int, ffi.Pointer, ffi.Pointer); # Example, adjust names
public static void _staticProgressCallback(int, int, double, ffi.Pointer); # Example, adjust names
}
-keepclassmembers class sd_image.SDImage { # Keep members of SDImage
public int width; # Keep specific fields accessed from native side (if any direct access)
public int height;
public int channel;
public ffi.Pointer data;
*; # Keep all members for safety, refine if needed later
}
-keepclassmembers class com.example.local_diffusion.stable_diffusion_processor.StableDiffusionProcessor { # Keep members if needed
*; # Keep all members for safety, refine if needed later
}
-keep,allowshrinking,allowobfuscation class com.example.local_diffusion.** { # Keep your main app package
<methods>;
<fields>;
<init>(...);
}
-keepattributes Signature
-keepattributes *Annotation*
-keep class dart.ffi.* { *; }
-keep class io.flutter.plugin.platform.* { *; }
```
</details>
- Disabled minification and resource shrinking
<details open><summary>build.gradle</summary>
```gradle
plugins {
id "com.android.application"
id "kotlin-android"
id "dev.flutter.flutter-gradle-plugin"
}
def localProperties = new Properties()
def localPropertiesFile = rootProject.file('local.properties')
if (localPropertiesFile.exists()) {
localPropertiesFile.withReader('UTF-8') { reader ->
localProperties.load(reader)
}
}
def flutterVersionCode = localProperties.getProperty('flutter.versionCode')
if (flutterVersionCode == null) {
flutterVersionCode = '1'
}
def flutterVersionName = localProperties.getProperty('flutter.versionName')
if (flutterVersionName == null) {
flutterVersionName = '1.0'
}
android {
namespace "com.example.local_diffusion"
compileSdkVersion flutter.compileSdkVersion
ndkVersion flutter.ndkVersion
compileOptions {
sourceCompatibility JavaVersion.VERSION_1_8
targetCompatibility JavaVersion.VERSION_1_8
}
kotlinOptions {
jvmTarget = '1.8'
}
sourceSets {
main.java.srcDirs += 'src/main/kotlin'
}
defaultConfig {
// TODO: Specify your own unique Application ID (https://developer.android.com/studio/build/application-id.html).
applicationId "com.example.local_diffusion"
// You can update the following values to match your application needs.
// For more information, see: https://docs.flutter.dev/deployment/android#reviewing-the-gradle-build-configuration.
minSdkVersion flutter.minSdkVersion
targetSdkVersion flutter.targetSdkVersion
versionCode flutterVersionCode.toInteger()
versionName flutterVersionName
}
buildTypes {
release {
// TODO: Add your own signing config for the release build.
// Signing with the debug keys for now, so `flutter run --release` works.
minifyEnabled false
shrinkResources false
signingConfig signingConfigs.debug
proguardFiles getDefaultProguardFile('proguard-android.txt'), 'proguard-rules.pro'
}
}
}
flutter {
source '../..'
}
dependencies {}
```
</details>
- Modified memory management approach
- The issue appears specific to release mode builds while handling FFI image data transfer
### Expected results
Image should be displayed and saved successfully after generation, same as in debug mode.
### Actual results
App crashes with SIGSEGV (signal 11) after image generation completes
`signal 11 (SIGSEGV), code 2 (SEGV_ACCERR), fault addr 0x0000007600aac080
pid: 17513, tid: 18005, name: DartWorker
`
### Code sample
<details open><summary>Code sample</summary>
```dart
if (result.address != 0) {
final image = result.cast<SDImage>().ref;
final width = (message['width'] as num).toInt();
final height = (message['height'] as num).toInt();
final channels = image.channel;
final totalBytes = width * height * channels;
final rgbaBytes = Uint8List(width * height * 4);
// Crash occurs during image data processing
}
```
</details>
### Screenshots or Video
_No response_
### Logs
_No response_
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[√] Flutter (Channel stable, 3.27.2, on Microsoft Windows [version 10.0.19045.5371], locale fr-FR)
[√] Windows Version (Installed version of Windows is version 10 or higher)
[!] Android toolchain - develop for Android devices (Android SDK version 35.0.0)
X cmdline-tools component is missing
Run `path/to/sdkmanager --install "cmdline-tools;latest"`
See https://developer.android.com/studio/command-line for more details.
X Android license status unknown.
Run `flutter doctor --android-licenses` to accept the SDK licenses.
See https://flutter.dev/to/windows-android-setup for more details.
[√] Chrome - develop for the web
[X] Visual Studio - develop Windows apps
X Visual Studio not installed; this is necessary to develop Windows apps.
Download at https://visualstudio.microsoft.com/downloads/.
Please install the "Desktop development with C++" workload, including all of its default components
[√] Android Studio (version 2023.1)
[√] Connected device (4 available)
[√] Network resources
! Doctor found issues in 2 categories.
```
</details>
| c: crash,platform-android,dependency: dart,will need additional triage | low | Critical |
2,806,379,552 | angular | Add field to ResourceLoaderParams to find out whether the loader is triggered for the first time | ### Which @angular/* package(s) are relevant/related to the feature request?
_No response_
### Description
Sometimes, it's necessary to determine whether the loader was triggered for the first time. Before 19.1.x, we could have checked `param.previous.status` for Idle, but now, this status is already set to `loading` initially.
### Proposed solution
Hence, I'd like to propose another way to achieve the same, e.g., a `first` flag (similar to the first flag in `ngOnChanges`).
### Alternatives considered
We could use a flag in userland. This is doable and can be hidden in a helper function wrapping the loader. However, I assume using a first flag is easier. | area: core,state: has PR,core: reactivity,cross-cutting: signals | low | Minor |
2,806,384,481 | pytorch | Is there a PyTorch version that can work properly on the Thor platform based on the Blackwell architecture? | Is there a PyTorch version that can work properly on the Thor platform based on the Blackwell architecture? I encounter many errors when compiling PyTorch source code on the Thor platform, and I don't know how to solve them. Is there any expert who can help me
cc @ptrblck @msaroufim @eqy | module: cuda,triaged | low | Critical |
2,806,403,723 | excalidraw | Feature: both a background color and a scratch fill type? | I've noticed that quite often I need to stack shapes on top of one another for some fancier constructions. I use the non-solid fill types to highlight various things, but given that their background is transparent (apart of the shading), stacking them leaks the thing underneath. My current solution is to have 2 of every shape. An "underlay" with a solid white color, and then my actual shape on top with the needed shading.
<img width="669" alt="Image" src="https://github.com/user-attachments/assets/08a28964-2eae-4247-a791-f17606115aeb" />
As seen on the above image, this does work, but it needs a lot of effort to duplicate all shapes and handle overlaps between them, so sometimes I end up forgetting or just not wanting to bother (see the outer red shape's 4 corners, each bleeding the black strokes from underneath).
Wondering if it would be an interesting idea to add a non-transparent option for the shaded fill type? I'm aware that this would all of a sudden require *2* background colors, so happy to be told no, just thought I'd bring it up as food for thought. | enhancement | low | Minor |
2,806,407,068 | ui | [bug]: Documentation of toast component for manual installation is wrong | ### Describe the bug
If you go to the official documentation: https://ui.shadcn.com/docs/components/toast
And select manual installation, the three snippets of code that you are supposed to copy and paste into your project are the same one. I guess it was a mistake when creating the page.
Could you please fix this and put the correct code for every file?
Thanks
### Affected component/components
Toast
### How to reproduce
If you go to the official documentation: https://ui.shadcn.com/docs/components/toast
And select manual installation, the three snippets of code that you are supposed to copy and paste into your project are the same one. I guess it was a mistake when creating the page.
### Codesandbox/StackBlitz link
_No response_
### Logs
```bash
```
### System Info
```bash
All machines, since it is the official website :)
```
### Before submitting
- [x] I've made research efforts and searched the documentation
- [x] I've searched for existing issues | bug | low | Critical |
2,806,414,251 | ollama | deepseek-r1 `qwen` variants use a new pre-tokenizer, which is not implemented in the llama.cpp version used | ### What is the issue?
The newly supported `deepseek-r1` model variants that have `distill-qwen` in the name use a new pre-tokenizer. Support for this has been added to the latest llama.cpp (not sure if the release version or just the latest commit on the main branch).
The backend llama.cpp that Ollama uses should be updated to support this, since the `default` pre-tokenizer is very different than the bespoke version.
### OS
Linux, macOS, Windows, Docker, WSL2
### GPU
_No response_
### CPU
_No response_
### Ollama version
0.5.7 | bug | low | Minor |
2,806,434,387 | pytorch | torch.backends.cudnn.flags use error when test | ### 🐛 Describe the bug
https://github.com/pytorch/pytorch/blob/main/test/nn/test_convolution.py#L3311
There is a problem with the use of cudnn.flags() here. The original purpose was to test the accuracy of allow_tf32 when it was turned on and off, but the call to cudnn.flags() causes allow_tf32 to always be True.
```python
@onlyCUDA
@tf32_on_and_off(0.005)
def test_ConvTranspose2d_size_1_kernel(self, device):
x_cpu = torch.randn(2, 3, 5, 5)
conv_cpu = torch.nn.ConvTranspose2d(3, 3, kernel_size=1)
y_cpu = conv_cpu(x_cpu)
y = torch.rand_like(y_cpu)
y_cpu.backward(y)
with cudnn.flags(enabled=False):
conv_cuda = torch.nn.ConvTranspose2d(3, 3, kernel_size=1).to(device)
conv_cuda.bias.data.copy_(conv_cpu.bias.data)
conv_cuda.weight.data.copy_(conv_cpu.weight.data)
y_cuda = conv_cuda(x_cpu.to(device))
y_cuda.backward(y.to(device))
self.assertEqual(y_cpu, y_cuda, atol=1e-5, rtol=0, exact_device=False)
```
### Versions
Regardless of the environment
cc @csarofeen @ptrblck @xwang233 @eqy | module: cudnn,module: convolution,triaged | low | Critical |
2,806,442,186 | ui | [bug]: Checkbox | ### Describe the bug
When using the [Tasks demo page](https://ui.shadcn.com/examples/tasks) with Firefox 134.0.2 on Windows 11, the checkbox move when you click on it.
See the video below:
https://github.com/user-attachments/assets/704cd821-0be9-4322-b2cf-35ce163a2ce7
The problem only reproduces on Firefox.
### Affected component/components
Checkbox
### How to reproduce
1. Go to [Tasks demo page](https://ui.shadcn.com/examples/tasks)
2. Click on any checkbox on the first column
3. Observe the checkbox moving up and down
### Codesandbox/StackBlitz link
_No response_
### Logs
```bash
```
### System Info
```bash
Firefox 134.0.2 on Windows 11
```
### Before submitting
- [x] I've made research efforts and searched the documentation
- [x] I've searched for existing issues | bug | low | Critical |
2,806,451,493 | storybook | [Bug]: can't get scss to load despite storybook does not crash | ### Describe the bug
I can't get the storybook to show scss styling, I've tried everything (with and without @storybook/preset-scss), but it still doesn't show styling coming from scss files - it does work with inline styling.
Storybook doesn't complain. Most webpackFinal config options without @storybook/preset-scss resulted with a storybook error, complaining about scss characters.
I have a react app, with webpack5, using scss&css.
```
"@storybook/addon-essentials": "^8.5.0",
"@storybook/addon-interactions": "^8.5.0",
"@storybook/addon-onboarding": "^8.5.0",
"@storybook/addon-webpack5-compiler-swc": "^2.0.0",
"@storybook/blocks": "^8.5.0",
"@storybook/preset-scss": "^1.0.3",
"@storybook/react": "^8.5.0",
"@storybook/react-webpack5": "^8.5.0",
"style-loader": "4.0.0",
"sass": "^1.81.0",
"sass-loader": "^16.0.4",
"css-loader": "6.11.0",
```
main.ts
```
const config: StorybookConfig = {
addons: [
'@storybook/addon-webpack5-compiler-swc',
'@storybook/addon-onboarding',
'@storybook/addon-essentials',
'@chromatic-com/storybook',
'@storybook/addon-interactions',
'@storybook/preset-scss',
],
core: {
builder: 'webpack5',
},
framework: {
name: '@storybook/react-webpack5',
options: {},
},
stories: [
'../src/**/*.mdx',
'../src/**/*.stories.@(js|jsx|mjs|ts|tsx)',
'../src/browser/storybook-utils/**/*.stories.@(js|jsx|mjs|ts|tsx)',
],
webpackFinal: async (config) => {
config.module.rules.push({
test: /\.scss$/,
use: [
'style-loader',
{
loader: 'css-loader',
options: {
importLoaders: 1, // Ensure sass-loader runs before css-loader
},
},
'sass-loader',
],
include: path.resolve(path.dirname(import.meta.url), '../'),
});
if (config.resolve) {
config.resolve.plugins = [
...(config.resolve.plugins || []),
new TsconfigPathsPlugin({
extensions: config.resolve.extensions,
}),
];
}
return config;
},
};
export default config;
```
```
import './ModalContainerHeader.scss';
...
<div className="ModalContainerHeader">
<div className="testOne" style={{ color: 'red' }}>
Test
</div>
<div className="testTwo">Test</div>
</div>
```
```
.ModalContainerHeader {
.testTwo {
color: red;
}
```
<img width="122" alt="Image" src="https://github.com/user-attachments/assets/36977036-52ed-4086-861b-f7462750fd32" />
<img width="408" alt="Image" src="https://github.com/user-attachments/assets/1f1b3d4c-3a55-49f1-9521-14adb084b481" />
any ideas what to do?
### Reproduction link
can't share sorry
### Reproduction steps
_No response_
### System
```bash
worked ok
```
### Additional context
_No response_ | bug,needs triage | low | Critical |
2,806,465,641 | neovim | gF should recognize more file/line/column patterns and seek to column | ### Problem
Right now the [gF](https://neovim.io/doc/user/editing.html#gF) feature seems to only support a limited set of patterns. In particular, it doesn't seem to work with python's backtrace format:
```
Traceback (most recent call last):
File "/path/to/file.py", line 123, in <module>
```
It also seems to completely ignore column info since gF on `file:123:12:` (eg as emitted by modern gcc and clang) seeks to the 123rd line but not the 12th column.
One of the things I miss most from vscode is that I can run a command in the terminal and just cmd-click in the output when it is referring to a file and go right where I need to go.
### Expected behavior
The vscode terminal's logic to detect file/line/column seems to work fairly well in practice, so it may be a good source of inspiration. It looks like [this](https://github.com/microsoft/vscode/blob/ce2c2f3c79a32b9917e32c61e058392dc5a1b6aa/src/vs/workbench/contrib/terminalContrib/links/browser/terminalLinkParsing.ts#L64-L126) is list of 3 regexes that they use, along with examples patterns that they are trying to match.
Once you have the parser able to recognize columns, it would be nice if `gF` automatically went to the right column as well.
Maybe this should be a new ticket, but it would be really nice if the built-in `:terminal` was automatically enhanced to recognize these patterns and A) apply a highlight group, possibly only when the cursor or mouse is over one (bonus points if the highlight group was only applied when the file exists!), and B) consider a default mapping on `<LeftMouse>` or `<C-LeftMouse>` to go to the file.
Another nice feature would be something to send the matching locations to the quickfix list. I have tried `:cgetbuffer` from the terminal and it does not work well, at least with the default `errorformat`. Among other issues, I wasn't able to make one that would match file:line:column in the middle of a line and not treat anything before the file as part of the message rather than the file. It would be especially cool if there was support for only sending the output from the last command to quickfix using OSC codes from the prompt to delimit command boundaries (at least kitty and vscode support the [OSC 133](https://gitlab.freedesktop.org/Per_Bothner/specifications/blob/master/proposals/semantic-prompts.md) A/B/C/D sequence for this which is also useful for #9209). This would let you run your build in a terminal and still get all the niceties like live progress reporting and colorful output, while still being able to use the quickfix list as if you used `:make` | jumps-navigation,editor | low | Critical |
2,806,472,718 | material-ui | [docs] Tabs need two clicks to work on Firefox | ### Related page
https://mui.com/material-ui/react-tabs/
### Kind of issue
Broken demo
### Issue description
For some examples It needs to click twice to change the tab. That's happening to the firefox browser in chromium based browsers looks fine.
### Context
Open the tabs demo documentation in Firefox. I'm on macOS. Going to click to change the tabs, and I'm doing a click on a different tab and looks like clicked, does an effect, stays to the current tab, click again, going to the tab that I clicked. After refresh, it worked, but I cannot consider a cache issue since it was the first time that I visited the page from this browser.
**Search keywords**: tabs, double click | docs,component: tabs,status: waiting for author,support: docs-feedback | low | Critical |
2,806,540,487 | vscode | HTML can't wite any more code Visual Studio Code |
Type: <b>Bug</b>
I have issue with Visual Studio Code. All of a sudden , i can't write code anymore. let me explain: when i try to mwrite code, it keeps inserting closing tag and i can no longer open any tags. The closing tags appear for all tags like DIV, MAIN, BODY. However ,i've checked my code many times, and it's perfectly fine. I also tried with other project , and i'm experiencing the same problem. I need to solve this issue because i work with VSC. Thank You!
VS Code version: Code 1.96.2 (Universal) (fabdb6a30b49f79a7aba0f2ad9df9b399473380f, 2024-12-19T10:22:47.216Z)
OS version: Darwin arm64 23.6.0
Modes:
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Apple M3 (8 x 2400)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: enabled_on<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_graphite: disabled_off<br>video_decode: enabled<br>video_encode: enabled<br>webgl: enabled<br>webgl2: enabled<br>webgpu: enabled<br>webnn: disabled_off|
|Load (avg)|5, 12, 7|
|Memory (System)|16.00GB (3.14GB free)|
|Process Argv|--crash-reporter-id 8d604037-1790-40a6-96c0-15e093dc2ea9|
|Screen Reader|no|
|VM|0%|
</details>Extensions: none<details>
<summary>A/B Experiments</summary>
```
vsliv368:30146709
vspor879:30202332
vspor708:30202333
vspor363:30204092
vscod805:30301674
binariesv615:30325510
vsaa593:30376534
py29gd2263:31024239
c4g48928:30535728
azure-dev_surveyone:30548225
962ge761:30959799
pythonnoceb:30805159
pythonmypyd1:30879173
2e7ec940:31000449
pythontbext0:30879054
cppperfnew:31000557
dsvsc020:30976470
pythonait:31006305
dsvsc021:30996838
dvdeprecation:31068756
dwnewjupyter:31046869
newcmakeconfigv2:31071590
nativerepl1:31139838
pythonrstrctxt:31112756
nativeloc1:31192215
cf971741:31144450
iacca1:31171482
notype1:31157159
5fd0e150:31155592
dwcopilot:31170013
stablechunks:31184530
6074i472:31201624
dwoutputs:31217127
hdaa2157:31222309
copilot_t_ci:31222730
```
</details>
<!-- generated by issue reporter --> | info-needed | low | Critical |
2,806,557,437 | ant-design | AutoComplete customize input size | ### Reproduction link
[](https://codesandbox.io/p/sandbox/customize-input-component-antd-5-23-2-forked-grw9hy?file=%2Fdemo.tsx)
### Steps to reproduce
Switch between small, middle and large
### What is expected?
AutoComplete wrapper should size to custom Input's height
### What is actually happening?
* When size prop is set only on the custom Input, AutoComplete wrapper itself keeps fixed height.
* When size prop is set on both Input and AutoComplete,then it works almost as it should, except at large size AutoComplete height is larger the Input's, also it shows the warning in console: Warning: [antd: AutoComplete] You need to control style self instead of setting `size` when using customize input.
| Environment | Info |
| --- | --- |
| antd | 5.23.2 |
| React | 18 |
| System | Windows |
| Browser | Firefox |
<!-- generated by ant-design-issue-helper. DO NOT REMOVE --> | improvement | low | Minor |
2,806,591,034 | storybook | [Bug]: “The current testing environment is not configured to support act” errors when changing between component docs pages | ### Describe the bug
When switching between component docs pages you get bunch of
```
The current testing environment is not configured to support act(...)
```
and occasionally
```
An update to ZoomElement inside a test was not wrapped in act(...).
```
errors in browser console.
### Reproduction link
https://stackblitz.com/edit/github-qczcaoih?file=package.json&preset=node
### Reproduction steps
1. Navigate to Button docs page
2. While on Button docs page, navigate to Header or Page docs page.
3. Check browser console for errors.
### System
```bash
System:
OS: Linux 5.0 undefined
CPU: (8) x64 Intel(R) Core(TM) i9-9880H CPU @ 2.30GHz
Shell: 1.0 - /bin/jsh
Binaries:
Node: 18.20.3 - /usr/local/bin/node
Yarn: 1.22.19 - /usr/local/bin/yarn
npm: 10.2.3 - /usr/local/bin/npm <----- active
pnpm: 8.15.6 - /usr/local/bin/pnpm
npmPackages:
@storybook/addon-essentials: ^8.6.0-alpha.1 => 8.6.0-alpha.1
@storybook/addon-interactions: ^8.6.0-alpha.1 => 8.6.0-alpha.1
@storybook/addon-onboarding: ^8.6.0-alpha.1 => 8.6.0-alpha.1
@storybook/addon-webpack5-compiler-swc: ^2.0.0 => 2.0.0
@storybook/blocks: ^8.6.0-alpha.1 => 8.6.0-alpha.1
@storybook/react: ^8.6.0-alpha.1 => 8.6.0-alpha.1
@storybook/react-webpack5: ^8.6.0-alpha.1 => 8.6.0-alpha.1
@storybook/test: ^8.6.0-alpha.1 => 8.6.0-alpha.1
storybook: ^8.6.0-alpha.1 => 8.6.0-alpha.1
```
### Additional context
React version: 19
Brower: Chrome, Firefox | bug,react,needs triage,upgrade:8.5 | low | Critical |
2,806,591,543 | rust | compiletest: investigate if it's possible to do run-time capability detection against *target* | Some tests are conditioned on *capabilities*, which are most accurately detected at run-time *against* the **target** under cross-compilation scenarios (compared to having to maintain allowlists or denylists of targets). However, implementing this is non-trivial:
- You have to build *and* run the capability run-time test as an executable on the **target**, *not* the **host**. This may involve having to work with custom test runners (e.g. `wasmtime`) or remote-test-server.
- The *target* may not support std! | E-hard,C-enhancement,A-cross,T-compiler,T-bootstrap,A-compiletest,E-needs-design | low | Minor |
2,806,596,749 | flutter | [ Web ] Flutter web Autofill doesn't work when browser password manager is protected with Windows Hello | ### Steps to reproduce
- Create a basic login form with a login and password fields, with autofill
- Activate Windows Hello protection on browser (in Chrome : password manager -> settings -> Activate Windows Hello)
- First run on browser -> save credentials
- Restart your web app
- Use this second run to use the saved credentials, with autofill
### Expected results
The selected credentials should fill the login/password form fields
### Actual results
- [x] if the browser uses its own password manager directly, the autofill works perfectly (tested with Chrome, Mozilla Firefox)
- [ ] if the browser uses Windows Hello to protect the password manager, the fields are not filled after Windows validation (tested only with Chrome, because I didnt success to make Mozilla Firefox ask for Windows Hello confirmation)
### Code sample
<details open><summary>Code sample</summary>
```dart
Form(
child: AutofillGroup(
child: Column(
mainAxisSize: MainAxisSize.min,
children: [
//...
TextField(
autofillHints: const [AutofillHints.username],
controller: loginCtrl,
),
TextField(
autofillHints: const [AutofillHints.password],
obscureText: true,
controller: pwdCtrl,
),
```
</details>
| a: text input,platform-windows,platform-web,has reproducible steps,browser: chrome-desktop,team-web,fyi-text-input,found in release: 3.27,found in release: 3.28 | low | Minor |
2,806,608,458 | kubernetes | Pass ctx to the remaining WaitForXXX methods in test/integration/util/util.go | ### What would you like to be added?
Pass ctx to the remaining WaitForXXX methods in test/integration/util/util.go
### Why is this needed?
We should use existing test context consistently. | sig/scheduling,kind/feature,needs-triage | low | Minor |
2,806,630,532 | flutter | [A11y]: Slider labels not included in the `Semantics` widget | ### Steps to reproduce
When creating `Slider` widgets, the `Slider.label` information isn't passed to the semantics widget.
### Expected results
Sliders should announce their labels when tabbing through with a screen reader such as NVDA. I can't test platforms other than Windows or Chrome while at work.
### Actual results
NVDA announces "Slider", then the percentage.
The third slider works perfectly. Looking at the implementation of the `Slider` widget, it seems that there is no `label` property being passed to the `Semantics` widget at the end of the `_buildMaterialSlider` method.
I've tested this locally, and adding the `label` argument fixes this.
Am I OK to submit a PR with this change? I don't really know how all of this works. I'm keen to help, but - depending on how much vetting I'll need - it might be quicker for someone else to add `label: widget.label` to line 1083 of `slider.dart`. 😄
Thank you for a wonderful framework!!
### Code sample
<details open><summary>Code sample</summary>
```dart
import 'package:flutter/material.dart';
import 'package:flutter/rendering.dart';
void main() {
runApp(const MyApp());
}
class MyApp extends StatelessWidget {
const MyApp({super.key});
// This widget is the root of your application.
@override
Widget build(BuildContext context) {
final renderBinding = RendererBinding.instance;
if (!renderBinding.semanticsEnabled) {
renderBinding.ensureSemantics();
}
return MaterialApp(
title: 'Flutter Demo',
theme: ThemeData(
colorScheme: ColorScheme.fromSeed(seedColor: Colors.deepPurple),
useMaterial3: true,
),
home: const MyHomePage(title: 'Flutter Demo Home Page'),
);
}
}
class MyHomePage extends StatefulWidget {
const MyHomePage({super.key, required this.title});
final String title;
@override
State<MyHomePage> createState() => _MyHomePageState();
}
class _MyHomePageState extends State<MyHomePage> {
int _counter = 0;
void _incrementCounter() {
setState(() {
_counter++;
});
}
@override
Widget build(BuildContext context) {
const divider = 100;
return Scaffold(
appBar: AppBar(
backgroundColor: Theme.of(context).colorScheme.inversePrimary,
title: Text(widget.title),
),
body: Center(
child: Column(
mainAxisAlignment: MainAxisAlignment.center,
children: <Widget>[
const Text(
'You have pushed the button this many times:',
),
Text('$_counter'),
Slider(
autofocus: true,
value: _counter / divider,
onChanged: (value) => setState(
() => _counter = (value * divider).round(),
),
label: 'Screen readers will not pick up this label',
),
Slider(
value: _counter / divider,
onChanged: (value) => setState(
() => _counter = (value * divider).round(),
),
label: 'Screen readers will not speak this label either',
semanticFormatterCallback: (value) =>
'Screen readers will show this value. Counter is $_counter',
),
MergeSemantics(
// I use `MergeSemantics` to prevent NVDA from recognising a group.
child: Semantics(
label: 'This works perfectly',
child: Slider(
value: _counter / divider,
onChanged: (value) => setState(
() => _counter = (value * divider).round(),
),
label:
'We just need to wrap the `Slider` widget in a `MergeSemantics` and a `Semantics` widget and everything will be fixed',
),
),
)
],
),
),
floatingActionButton: FloatingActionButton(
onPressed: _incrementCounter,
tooltip: 'Increment',
child: const Icon(Icons.add),
),
);
}
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
[Upload media here]
</details>
### Logs
<details open><summary>Logs</summary>
```console
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[Γ£ô] Flutter (Channel master, 3.28.0-2.0.pre.38833, on Microsoft Windows [Version 10.0.26100.2894], locale en-GB) [2.5s]
ΓÇó Flutter version 3.28.0-2.0.pre.38833 on channel master at C:\Users\chris\flutter
ΓÇó Upstream repository https://github.com/flutter/flutter
ΓÇó Framework revision 97ca57cf08 (6 hours ago), 2025-01-23 06:38:20 +0100
ΓÇó Engine revision 97ca57cf08
ΓÇó Dart version 3.8.0 (build 3.8.0-24.0.dev)
ΓÇó DevTools version 2.42.0
[Γ£ô] Windows Version (11 Home 64-bit, 24H2, 2009) [4.1s]
[Γ£ô] Android toolchain - develop for Android devices (Android SDK version 34.0.0) [4.4s]
ΓÇó Android SDK at C:\Users\chris\AppData\Local\Android\sdk
ΓÇó Platform android-35, build-tools 34.0.0
ΓÇó Java binary at: C:\Program Files\Android\Android Studio\jbr\bin\java
This is the JDK bundled with the latest Android Studio installation on this machine.
To manually set the JDK path, use: `flutter config --jdk-dir="path/to/jdk"`.
ΓÇó Java version OpenJDK Runtime Environment (build 17.0.6+0-b2043.56-9586694)
ΓÇó All Android licenses accepted.
[Γ£ô] Chrome - develop for the web [119ms]
ΓÇó Chrome at C:\Program Files\Google\Chrome\Application\chrome.exe
[Γ£ô] Visual Studio - develop Windows apps (Visual Studio Community 2022 17.12.4) [117ms]
ΓÇó Visual Studio at C:\Program Files\Microsoft Visual Studio\2022\Community
ΓÇó Visual Studio Community 2022 version 17.12.35707.178
ΓÇó Windows 10 SDK version 10.0.22621.0
[Γ£ô] Android Studio (version 2022.2) [16ms]
ΓÇó Android Studio at C:\Program Files\Android\Android Studio
ΓÇó Flutter plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/9212-flutter
ΓÇó Dart plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/6351-dart
ΓÇó Java version OpenJDK Runtime Environment (build 17.0.6+0-b2043.56-9586694)
[Γ£ô] VS Code, 64-bit edition (version 1.96.2) [14ms]
ΓÇó VS Code at C:\Program Files\Microsoft VS Code
ΓÇó Flutter extension version 3.102.0
[Γ£ô] Connected device (3 available) [290ms]
ΓÇó Windows (desktop) ΓÇó windows ΓÇó windows-x64 ΓÇó Microsoft Windows [Version 10.0.26100.2894]
ΓÇó Chrome (web) ΓÇó chrome ΓÇó web-javascript ΓÇó Google Chrome 131.0.6778.265
ΓÇó Edge (web) ΓÇó edge ΓÇó web-javascript ΓÇó Microsoft Edge 132.0.2957.115
[Γ£ô] Network resources [412ms]
ΓÇó All expected network resources are available.
ΓÇó No issues found!
```
</details>
| in triage | low | Major |
2,806,638,283 | react | Bug: onChange Not Triggering for range input After Dynamic Updates | <!--
Please provide a clear and concise description of what the bug is. Include
screenshots if needed. Please test using the latest version of the relevant
React packages to make sure your issue has not already been fixed.
-->
## Description
When using range input in React, the onChange event does not trigger properly under certain dynamic updates to its step and max attributes. Specifically, if the step and max values are updated dynamically, and the range input value is reset to 0, subsequent user interactions with the range slider fail to trigger the onChange event as expected. This behavior does not occur in a pure JavaScript implementation.
**React version:** 19.0.0
## Steps To Reproduce
1. Open the StackBlitz example: [React example](https://stackblitz.com/edit/vitejs-vite-srhuep2z?file=src%2FApp.jsx)
2. Click the middle of the range input, changing its value to `500`.
3. Click the **"Change to step 0.5 max 1"** button, observe that the range input is set to `0` (expected behavior).
4. Click the **"Change to step 500 max 1000"** button, observe that the range input is set back to `0`. Then click the middle of the range input.
## The current behavior
After completing step 4, the value of the range input remains at 0, even though the user clicks on the middle of the range slider. The onChange event is not triggered, and the value does not update.
## The expected behavior
After completing step 4, the value of the range input should update to 500 when the user clicks on the middle of the range slider. The onChange event should be triggered correctly, reflecting the value change.
## Link to code example
- [React implementation reproducing the issue](https://stackblitz.com/edit/vitejs-vite-srhuep2z?file=src%2FApp.jsx)
- [Pure JavaScript implementation without the issue](https://codepen.io/wrxue/pen/emOPrbL)
---
https://github.com/user-attachments/assets/0d471278-8b12-499d-8b86-819d03a17d2f | Status: Unconfirmed | medium | Critical |
2,806,647,618 | react | [DevTools Bug]: [DevTools Bug] Cannot reorder children for node "0" because no matching node was found in the Store. | ### Website or app
https://www.autotrack.nl/aanbod
### Repro steps
1. Start profiling in chrome
2. Click on any filter
3. Uncaught Error: Cannot reorder children for node "0" because no matching node was found in the Store.
Dismiss
The error was thrown at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1175534
at v.emit (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1140783)
at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1142390
at bridgeListener (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1552529)
### How often does this bug happen?
Every time
### DevTools package (automated)
react-devtools-extensions
### DevTools version (automated)
6.0.1-c7c68ef842
### Error message (automated)
Cannot reorder children for node "0" because no matching node was found in the Store.
### Error call stack (automated)
```text
at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1175534
at v.emit (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1140783)
at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1142390
at bridgeListener (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:1:1552529)
```
### Error component stack (automated)
```text
```
### GitHub query string (automated)
```text
https://api.github.com/search/issues?q=Cannot reorder children for node because no matching node was found in the Store. in:title is:issue is:open is:public label:"Component: Developer Tools" repo:facebook/react
``` | Type: Bug,Status: Unconfirmed,Component: Developer Tools | medium | Critical |
2,806,652,646 | rust | pthread_exit crashes on threads created by std::thread::spawn in 1.84, not 1.83, breaking pyo3-log | ## Meta
Tested on Ubuntu 24.04 and Amazon Linux 2, x86_64.
## Workaround to the production problem
In your Cargo.toml that is compiling your Python module, set
```toml
[profile.release]
debug = 0
lto = false
```
This prevents the 1.84 crashes.
However, there is still UB going on even with this setting.
## The Production Problem
The crate [pyo3-log](https://docs.rs/pyo3-log/0.12.1/pyo3_log/) installs a bridge that makes [log](https://docs.rs/log/latest/log/trait.Log.html) functions call into Python. This means that all calls to `logging::info!` etc will take the GIL.
Python has a stage during interpreter shutdown where attempts to take the GIL will cause a `pthread_exit`. Python 3.14 (still Alpha today, targeted to be released by the end of this year) will change this in python/cpython#87135 - but that will take some time to reach people.
This means that if you have a Python program that uses a Rust library and pyo3-log, that spawning a Rust thread, that is calling `logging::info!` in a way unsynchronized with interpreter exit, you'll have unpredicatable crashes in 1.84.
## Minified Program
This program:
```rust
use std::ffi::c_void;
extern "C" {
fn pthread_exit(retval: *const c_void);
}
fn main() {
std::thread::spawn(|| {
unsafe { pthread_exit(std::ptr::null()); }
});
std::thread::sleep(std::time::Duration::from_secs(1));
}
```
when compiled with the following options
```
rustc +1.84 d.rs -Cpanic=abort -Cdebuginfo=limited
```
crashes with this confusing error
```
thread '<unnamed>' panicked at core/src/panicking.rs:223:5:
panic in a function that cannot unwind
stack backtrace:
0: rust_begin_unwind
at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/std/src/panicking.rs:665:5
1: core::panicking::panic_nounwind_fmt::runtime
at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/core/src/panicking.rs:119:22
2: core::panicking::panic_nounwind_fmt
at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/core/src/intrinsics/mod.rs:3535:9
3: core::panicking::panic_nounwind
at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/core/src/panicking.rs:223:5
4: core::panicking::panic_cannot_unwind
at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/core/src/panicking.rs:315:5
5: std::sys::pal::unix::thread::Thread::new::thread_start
at /rustc/9fc6b43126469e3858e2fe86cafb4f0fd5068869/library/std/src/sys/pal/unix/thread.rs:99:9
6: start_thread
7: clone
note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.
thread caused non-unwinding panic. aborting.
Aborted
```
This crash happens:
1. Only on 1.84, not on 1.83
2. Only when debuginfo is enabled, but *even if the binary is stripped*.
When using `-C panic=unwind` instead, on all versions of the compiler, you get this error:
```
FATAL: exception not rethrown
Aborted (core dumped)
```
I seen the claim in Zulip (https://rust-lang.zulipchat.com/#narrow/channel/122651-general/topic/pthread_exit.20from.20a.20Rust-spawned.20thread) that this is undefined behavior, but I'll rather not break pyo3-log | T-compiler,C-bug,T-libs,A-thread,S-has-mcve,S-has-bisection | low | Critical |
2,806,660,658 | kubernetes | Replace reflect.DeepEqual with cmp.Diff in pkg/scheduler tests | ### What would you like to be added?
In some tests, `reflect.DeepEqual` is used to compare expected and actual values. It should be replaced to `cmp.Diff` if possible.
### Why is this needed?
According to the [SIG Scheduling guidelines](https://github.com/kubernetes/community/blob/master/sig-scheduling/CONTRIBUTING.md#technical-and-style-guidelines), `cmp.Diff` should be used instead of `reflect.DeepEqual`. | sig/scheduling,kind/feature,needs-triage | low | Minor |
2,806,663,453 | PowerToys | Error after enabling high contrast(accesibility panel) from default unlock window | ### Microsoft PowerToys version
0.87.0.0
### Installation method
GitHub
### Running as admin
None
### Area(s) with issue?
General
### Steps to reproduce
lock pc and enable high contrast throw accessibility panel
[2025-01-23.txt](https://github.com/user-attachments/files/18519748/2025-01-23.txt)
### ✔️ Expected Behavior
No errors on this behavior
### ❌ Actual Behavior
Error window displayed
### Other Software
_No response_ | Issue-Bug,Needs-Triage | low | Critical |
2,806,669,948 | kubernetes | [Flaking Test] [sig-apps] Kubernetes e2e suite.[It] [sig-apps] ReplicaSet Replace and Patch tests [Conformance] | ### Which jobs are flaking?
master-blocking
- gce-cos-master-default
### Which tests are flaking?
Kubernetes e2e suite.[It] [sig-apps] ReplicaSet Replace and Patch tests [Conformance]
[Prow](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-gci-gce/1882170659614756864)
[Triage](https://storage.googleapis.com/k8s-triage/index.html?test=ReplicaSet%20Replace%20and%20Patch%20tests%20&xjob=e2e-kops)
### Since when has it been flaking?
[1/10/2025, 5:10:10 AM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-kind-conformance-parallel/1877628702888562688)
[1/10/2025, 11:40:02 AM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-gci-gce-alpha-enabled-default/1877727102057320448)
[1/19/2025, 6:47:20 AM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/e2e-ci-kubernetes-e2e-al2023-aws-conformance-canary/1880914873404100608)
[1/20/2025, 7:25:30 PM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-ubuntu-ec2-containerd/1881467771536019456)
[1/22/2025, 5:57:09 PM](https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-gci-gce/1882170659614756864)
### Testgrid link
https://testgrid.k8s.io/sig-release-master-blocking#gce-cos-master-default
### Reason for failure (if possible)
```
{ failed [FAILED] failed to see replicas of test-rs in namespace replicaset-5368 scale to requested amount of 3: watch closed before UntilWithoutRetry timeout
In [It] at: k8s.io/kubernetes/test/e2e/apps/replica_set.go:552 @ 01/22/25 21:08:36.337
}
```
### Anything else we need to know?
N/A
### Relevant SIG(s)
/sig apps
cc: @kubernetes/release-team-release-signal | kind/flake,sig/apps,needs-triage | low | Critical |
2,806,699,847 | PowerToys | Your message on my computer: A problem has occurred - Report the error in the GitHub repository for PowerToys - 2025-01-23 | ### Microsoft PowerToys version
0.87.1.0
### Installation method
WinGet
### Running as admin
None
### Area(s) with issue?
Settings
### Steps to reproduce
Date: 23.01.2025 11:34:34
Exception:
System.Windows.Markup.XamlParseException: Set property
[2025-01-23.txt](https://github.com/user-attachments/files/18519964/2025-01-23.txt)
### ✔️ Expected Behavior
_No response_
### ❌ Actual Behavior
?
### Other Software
_No response_ | Issue-Bug,Needs-Triage | low | Critical |
2,806,712,685 | transformers | Paliegemma Pad Token not Masked | ### System Info
- `transformers` version: 4.48.0
- Platform: Linux-6.1.92-3-x86_64-with-glibc2.31
- Python version: 3.10.15
- Huggingface_hub version: 0.26.2
- Safetensors version: 0.4.5
- Accelerate version: 1.1.1
- Accelerate config: not found
- PyTorch version (GPU?): 2.5.1+cu124 (True)
- Tensorflow version (GPU?): not installed (NA)
- Flax version (CPU?/GPU?/TPU?): not installed (NA)
- Jax version: not installed
- JaxLib version: not installed
- Using distributed or parallel set-up in script?: parallel set-up
- Using GPU in script?: Yes
- GPU type: NVIDIA A100 80GB PCIe
### Who can help?
@amyeroberts @molbap
**Description:**
I encountered an unexpected behavior while using the Paligemma model. Specifically, the pad tokens' attention masks are being processed to zero (i.e., unmasked status) during the execution of the `self._update_causal_mask(attention_mask, token_type_ids, past_key_values, cache_position, input_ids, inputs_embeds, is_training)` function.
After investigating, I found that this behavior stems from a change in the code between the latest version (4.48.0) and the previous version (e.g., 4.43.0). The order in which the pad tokens are masked has been reversed, leading to inconsistent behavior, especially when using left padding.
### Code Comparison
**Latest Version (4.48.0):**
```python
if attention_mask is not None:
causal_mask = causal_mask.clone() # copy to contiguous memory for in-place edit
mask_length = attention_mask.shape[-1]
padding_mask = causal_mask[:, :, :, :mask_length] + attention_mask[:, None, None, :].to(causal_mask.device)
padding_mask = padding_mask == 0
causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill(
padding_mask, min_dtype
)
# we are training thus we need to create a full mask on the image + prefix but causal on suffix
if is_training:
causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill(
token_type_ids[:, None, None, :].to(causal_mask.device) == 0, 0
)
```
**Previous Version (4.43.0):**
```python
if token_type_ids is not None and labels is not None:
# we are training thus we need to create a full mask on the image + prefix but causal on suffix
target_length = cache_position[-1] + 1
causal_mask = torch.full(
(sequence_length, target_length), fill_value=min_dtype, dtype=dtype, device=device
)
if sequence_length != 1:
causal_mask = torch.triu(causal_mask, diagonal=1)
causal_mask *= torch.arange(target_length, device=device) > cache_position.reshape(-1, 1)
causal_mask = causal_mask[None, None, :, :].expand(inputs_embeds.shape[0], 1, -1, -1)
if attention_mask is not None:
causal_mask = causal_mask.clone() # copy to contiguous memory for in-place edit
mask_length = attention_mask.shape[-1]
padding_mask = causal_mask[:, :, :, :mask_length] + attention_mask[:, None, None, :].to(
causal_mask.device
)
# unmask the prefill
causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill(
token_type_ids[:, None, None, :].to(causal_mask.device) == 0, 0
)
padding_mask = padding_mask == 0
causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill(
padding_mask, min_dtype
)
```
**Issue Details**
In the latest version (4.48.0), the pad tokens are first masked using the padding_mask:
```python
padding_mask = padding_mask == 0
causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill(
padding_mask, min_dtype
)
```
However, when the following block of code is executed during training:
```python
if is_training:
causal_mask[:, :, :, :mask_length] = causal_mask[:, :, :, :mask_length].masked_fill(
token_type_ids[:, None, None, :].to(causal_mask.device) == 0, 0
)
```
the pad tokens (which have token_type_ids == 0) are unmasked again, effectively reverting the previous masking operation.
**Impact**
This behavior is particularly problematic when using left padding, as the pad tokens end up being unmasked in the latest version, whereas they were correctly masked in the previous version (4.43.0).
**Question**
Is this an intentional change, or is it a bug? If it’s a bug, could you please provide guidance on how to address it? I’d be happy to help with a fix if needed.
Thanks for your support!
### Information
- [x] The official example scripts
- [ ] My own modified scripts
### Tasks
- [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...)
- [x] My own task or dataset (give details below)
### Reproduction
```python
from PIL import Image
from transformers import AutoProcessor, PaliGemmaForConditionalGeneration
processor = AutoProcessor.from_pretrained("path")
model = PaliGemmaForConditionalGeneration.from_pretrained("path")
prompt = ["<image>caption en","<image>caption en wwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww wwwwwwwwwwwwwwwwwww"]
labels = ["123123123123","456456456456"]
image_file = ["image path"] * 2
raw_image = [Image.open(i).convert("RGB") for i in image_file]
inputs = processor(text=prompt, images=raw_image, suffix=labels,
return_tensors="pt", padding="longest")
model(**inputs)
```
### Expected behavior
The pad tokens end up being unmasked. | bug,Multimodal,VLM | low | Critical |
2,806,743,430 | godot | ColorPicker load/save swatches button issues | ### Tested versions
v4.4.beta1.official [d33da79d3]
### System information
Fedora Linux 40 (KDE Plasma) on Wayland - Wayland display driver, Single-window
### Issue description
ColorPicker load/save swatches button moves when clicked:
https://github.com/user-attachments/assets/e16bfa09-8fb9-492f-acb0-44ae6a18f443
The PopupMenu is also not behaving the same as other Popupmenus in the ColorPicker in RtL:
https://github.com/user-attachments/assets/b9a7b0cf-89d0-454a-955f-0d5997194274
There is also an issue in single window mode, where you can't close the menu by clicking on the button again:
https://github.com/user-attachments/assets/73ee9d81-e902-4d8a-9c38-975f2da58e21
### Steps to reproduce
Try any ColorPicker in the editor
### Minimal reproduction project (MRP)
N/A | bug,topic:gui | low | Minor |
2,806,759,283 | flutter | revert deprecating Color withOpacity / opacity | Read through the [code](https://github.com/flutter/flutter/blob/master/engine/src/flutter/lib/ui/painting.dart#L328-L332) and [migration guide](https://docs.flutter.dev/release/breaking-changes/wide-gamut-framework) but still having trouble supporting the end result here.
### Axioms
1. the vast majority of production codebases use .withOpacity(x), where x is 0.0–1.0 [1](https://github.com/search?type=code&q=.withOpacity%28)
1. the internal implementation is now .withOpacity(x) => .withAlpha((255.0*x).round()); [2](https://github.com/flutter/flutter/blob/master/engine/src/flutter/lib/ui/painting.dart#L328-L332)
1. the migration docs say to migrate .withOpacity(x) => .withValues(alpha: x); [3](https://docs.flutter.dev/release/breaking-changes/wide-gamut-framework#migrate-withopacity)
1. .withValues(alpha: x) is more verbose and less obvious than .withOpacity(x) [4](https://x.com/spiffyeehaw/status/1882459279564886111?s=46)
1. **designers communicate opacity as a percentage** from 0%–100% not an integer from 0–255 [5](https://help.figma.com/hc/en-us/articles/360041098433-Adjust-the-properties-of-an-image#:~:text=Click%20the%20thumbnail%20to%20open,by%20clicking%20the%20eye%20icon.)
### Questions
1. why doesn't the internal implementation match the migration docs
1. why does withAlpha(x) take an int 0–255 and withValues(alpha: x) take a double 0–1.0
1. why does .withOpacity need to be deprecated at all
### Recommendations
1. unify alpha to a double or an int across all color methods
1. **undeprecate withOpacity** method
1. **undeprecate opacity** getter | engine,customer: crowd,c: proposal,P3,team-engine,triaged-engine | high | Critical |
2,806,782,186 | material-ui | [Chip] Clickable area expands with `line-height` | ### Steps to reproduce
Steps:
1. Open this link to live example: https://stackblitz.com/edit/react-ryuspuxm?file=Demo.tsx
2. Click anywhere inside the black border box
### Current behavior
The chip is clicked even though the pointer is outside the Chip.
### Expected behavior
The chip should only be clicked when the pointer is inside the Chip.
### Context
Found in https://github.com/mui/mui-x/issues/16304
### Your environment
<details>
<summary><code>npx @mui/envinfo</code></summary>
```
System:
OS: macOS 14.5
Binaries:
Node: 22.2.0 - ~/.nvm/versions/node/v22.2.0/bin/node
npm: 10.7.0 - ~/.nvm/versions/node/v22.2.0/bin/npm
pnpm: 9.15.4 - ~/.nvm/versions/node/v22.2.0/bin/pnpm
Browsers:
Chrome: 131.0.6778.265
Edge: 132.0.2957.115
Safari: 17.5
npmPackages:
@mui/material: 6.4.1
```
</details>
**Search keywords**: Chip line-height | bug 🐛,component: chip,ready to take | low | Minor |
2,806,782,948 | flutter | iOS Zoom is very sensitive | ### Steps to reproduce
1. Open the camera on an iOS device
2. Pinch zoom
### Expected results
Zooming in using the pinch gesture functions the same as the native camera.
### Actual results
Zooming in using the pinch gesture is extremely sensitive & zooms in a lot, very quickly.
### Code sample
<details open><summary>Code sample</summary>
```dart
class CameraZoom {
final double minimunCameraZoom;
final double maximumCameraZoom;
final double currentCameraZoom;
final double zoomScale;
CameraZoom({
required this.minimunCameraZoom,
required this.maximumCameraZoom,
required this.currentCameraZoom,
required this.zoomScale,
});
bool isZoomOut() {
return zoomScale < 1.0;
}
double calculateZoomOut() {
final zoomToApply = zoomScale < 1.0
? currentCameraZoom - (maximumCameraZoom / 200)
: currentCameraZoom;
return zoomToApply;
}
double calculateZoomIn() {
final zoomToApply = zoomScale > 1.0
? currentCameraZoom + (maximumCameraZoom / 200)
: currentCameraZoom;
return zoomToApply;
}
double calculateZoomToApply() {
double scale = isZoomOut() ? calculateZoomOut() : calculateZoomIn();
if (scale <= minimunCameraZoom) {
scale = minimunCameraZoom;
} else if (scale >= maximumCameraZoom) {
scale = maximumCameraZoom;
}
return scale;
}
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
https://github.com/user-attachments/assets/4c0f94de-2daf-45be-9b89-95abf9f86b36
</details>
### Logs
<details open><summary>Logs</summary>
```console
[Paste your logs here]
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[Paste your output here]
```
</details>
| waiting for customer response,in triage | low | Minor |
2,806,796,663 | flutter | [iOS][PlatformView] Conflict between Flutter ListView scrolling and PlatformView | ### Steps to reproduce
This example highlights a conflict between Flutter's scrolling and the native platform view's `TextView` scrolling and text selection. Specifically, text selection in the TextView only works when `ListView` scrolling is disabled using physics: NeverScrollableScrollPhysics(). This limitation affects the usability of TextView and could also impact other types of platform views.
**Note:** This issue occurs only on iOS.
### Expected results
Users should be able to select text even when the TextView is inside a scrollable widget.
### Actual results
User cannot select text when TextView is inside a scrollable Widget and scrolling is enabled.
### Code sample
[flutter_native_text_view_sample_main.zip](https://github.com/user-attachments/files/18545585/flutter_native_text_view_sample_main.zip)
<details open><summary>Code sample</summary>
**Run this ready code:**
Screen has 3 Texts elements:
- the first one is flutter widget
- the other two are platform views.
Try to select text, it works with flutter widget but doesn't work with platform views. If scrolling is disable, text selection works also with platform views. After testing we think there is a conflict at low level.
Sample here:
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
- Flutter's SelectableText: Text selection works as expected.
- Native Platform View: Text selection is hindered by scrolling functionality.
https://github.com/user-attachments/assets/f7abf303-97b7-4d74-89cc-5ce17426dcae
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```Doctor summary (to see all details, run flutter doctor -v):
[!] Flutter (Channel [user-branch], 3.24.5, on macOS 14.2.1 23C71 darwin-arm64, locale en-IT)
! Flutter version 3.24.5 on channel [user-branch] at /Users/delfme/Development/flutter
Currently on an unknown channel. Run `flutter channel` to switch to an official channel.
If that doesn't fix the issue, reinstall Flutter by following instructions at https://flutter.dev/setup.
! Upstream repository unknown source is not a standard remote.
Set environment variable "FLUTTER_GIT_URL" to unknown source to dismiss this error.
[✓] Android toolchain - develop for Android devices (Android SDK version 34.0.0)
[✓] Xcode - develop for iOS and macOS (Xcode 15.2)
[✓] Chrome - develop for the web
[✓] Android Studio (version 2024.1)
[✓] Connected device (4 available)
! Error: Browsing on the local area network for iPhone. Ensure the device is unlocked and attached with a cable or associated with the same local area
network as this Mac.
The device must be opted into Developer Mode to connect wirelessly. (code -27)
[✓] Network resources
```
</details>
| platform-ios,framework,f: scrolling,a: platform-views,has reproducible steps,team-ios,found in release: 3.27,found in release: 3.29 | low | Critical |
2,806,796,688 | node | [v20.x] `parallel/test-buffer-tostring-range` fails on SmartOS | ### Test
parallel/test-buffer-tostring-range
### Platform
SmartOS
### Console output
```console
03:13:42 not ok 177 parallel/test-buffer-tostring-range
03:13:42 ---
03:13:43 duration_ms: 5550.52900
03:13:43 severity: fail
03:13:43 exitcode: 1
03:13:43 stack: |-
03:13:43 node:internal/buffer:961
03:13:43 super(bufferOrLength, byteOffset, length);
03:13:43 ^
03:13:43
03:13:43 RangeError: Array buffer allocation failed
03:13:43 at new ArrayBuffer (<anonymous>)
03:13:43 at new Uint8Array (<anonymous>)
03:13:43 at new FastBuffer (node:internal/buffer:961:5)
03:13:43 at Function.alloc (node:buffer:395:10)
03:13:43 at Object.<anonymous> (/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos23-x64/test/parallel/test-buffer-tostring-range.js:107:28)
03:13:43 at Module._compile (node:internal/modules/cjs/loader:1468:14)
03:13:43 at Module._extensions..js (node:internal/modules/cjs/loader:1547:10)
03:13:43 at Module.load (node:internal/modules/cjs/loader:1287:32)
03:13:43 at Module._load (node:internal/modules/cjs/loader:1103:12)
03:13:43 at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:180:12)
03:13:43
03:13:43 Node.js v20.18.3
03:13:43 ...
```
### Build links
- https://ci.nodejs.org/job/node-test-commit-smartos/58746/console
### Additional information
It's blocking v20 release https://github.com/nodejs/node/pull/56699
@nodejs/platform-smartos | smartos,flaky-test,v20.x | low | Critical |
2,806,815,184 | vscode | Braces colouring is off inside Dart multiline strings | <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
- VS Code Version: 1.96.4
- OS Version: Windows_NT x64 10.0.22631
Steps to Reproduce:
1. Open an empty folder with nothing inside
2. Disable all extensions
3. Add a new `file.dart` file
4. Write the following to it:
```dart
var myString = '''
Here is a string with ()
and some other examples {like (this)}.
''';
```
5. See the braces colouring (possibly, more on that later)

6. Change the string to be raw by adding `r` before the opening `'''`:

I've asked @DanTup to verify this but with the same VS Code version he could not see this behaviour on his machine.
If I close the editor (file opened) and reopen it (even removing `r` now) I see the braces colouring flash and go back to string colour.
His guess is that there may be some other racing here as https://github.com/microsoft/vscode/issues/230498. Same machine, a Samsung Galaxy Book Pro. To note we had problems with this colouring previously at https://github.com/microsoft/vscode/issues/209435 for Dart and we still have this opened for Python https://github.com/microsoft/vscode/issues/225179.
I can't reliably reproduce this (besides flashing) again on this _exact_ folder/file (unfortunatelly). But I saw this behaviour previously on Dart projects and this is an empty folder with only that file inside. | triage-needed | low | Critical |
2,806,817,554 | react-native | SectionList onViewableItemsChanged behaviour difference between iOS and Android/Web | ### Description
**Expected:**
`SectionList` `onViewableItemsChanged` should correctly report `viewableItems` across multiple sections when items from those sections are within the viewport. This is the current behaviour on Android and Web.
**Actual:**
`SectionList` `onViewableItemsChanged` will only report `viewableItems` for one section at at time, even if items from multiple sections are within the viewport. This is the current behaviour on iOS.
Notes:
* Using `getItemLayout` restores the expected behaviour on iOS.
* The reproducer uses the `viewableItems` parameter. However, the issue is also true of the `changed` parameter.
* I've tested the scenarios with and without passing varying `viewabilityConfig`.
* Not the same behaviour as reported in https://github.com/facebook/react-native/issues/48351. The following statement from that issue is not true of this issue: "If I remove getItemLayout entirely, the logic works fine regardless."
### Steps to reproduce
1. Open the supplied Expo reproducer
2. Scroll through the list
3. Observe the logs (messages list the sections with viewable items)
4. Do the above across multiple platforms, when scrolled to a point where multiple sections are within the viewport:
1. iOS will incorrectly only report a single section.
2. Android and Web will correctly report multiple sections.
Reproduced locally in vanilla `[email protected]` with new architecture. Also reproduced locally in `[email protected]` without new architecture.
### React Native Version
0.76.3
### Affected Platforms
Runtime - iOS
### Output of `npx react-native info`
```text
System:
OS: macOS 15.2
CPU: (12) arm64 Apple M2 Pro
Memory: 139.78 MB / 32.00 GB
Shell:
version: "5.9"
path: /bin/zsh
Binaries:
Node:
version: 20.12.2
path: ~/.nvm/versions/node/v20.12.2/bin/node
Yarn:
version: 1.22.19
path: /opt/homebrew/bin/yarn
npm:
version: 10.5.0
path: ~/.nvm/versions/node/v20.12.2/bin/npm
Watchman: Not Found
Managers:
CocoaPods:
version: 1.16.2
path: /Users/philipbulley/.rbenv/shims/pod
SDKs:
iOS SDK:
Platforms:
- DriverKit 24.2
- iOS 18.2
- macOS 15.2
- tvOS 18.2
- visionOS 2.2
- watchOS 11.2
Android SDK: Not Found
IDEs:
Android Studio: 2024.2 AI-242.23339.11.2421.12700392
Xcode:
version: 16.2/16C5032a
path: /usr/bin/xcodebuild
Languages:
Java:
version: 17.0.10
path: /usr/bin/javac
Ruby:
version: 3.3.0
path: /Users/philipbulley/.rbenv/shims/ruby
npmPackages:
"@react-native-community/cli":
installed: 15.0.1
wanted: 15.0.1
react:
installed: 18.3.1
wanted: 18.3.1
react-native:
installed: 0.76.3
wanted: 0.76.3
react-native-macos: Not Found
npmGlobalPackages:
"*react-native*": Not Found
Android:
hermesEnabled: true
newArchEnabled: true
iOS:
hermesEnabled: true
newArchEnabled: true
```
### Stacktrace or Logs
```text
n/a
```
### Reproducer
https://snack.expo.dev/@pbcodes/onviewableitemschanged-behaviour-difference-between-ios-and-android-or-web
### Screenshots and Videos
### iOS ❌
Scroll until items from 2 sections are within viewport
<img width="1366" alt="Image" src="https://github.com/user-attachments/assets/117fdcbf-b4bf-4003-b560-7659b9d8d177" />
Scroll further until the first section has left the viewport
<img width="1366" alt="Image" src="https://github.com/user-attachments/assets/a3a970f3-9198-4adc-9f8c-f47dc5401509" />
### Android ✅
Scroll until items from 2 sections are within viewport
<img width="1366" alt="Image" src="https://github.com/user-attachments/assets/e2bf13ec-2069-4de5-81e6-a8a499b34ba3" />
### Web ✅
Scroll until items from 2 sections are within viewport
<img width="1366" alt="Image" src="https://github.com/user-attachments/assets/043bc864-90d5-4226-b242-01a919c7fa33" /> | Platform: iOS,Platform: Android,Component: SectionList | low | Minor |
2,806,829,033 | pytorch | [custom ops] [2.7 nightly] custom ops with typing.List breaks when importing annotations from future | ### 🐛 Describe the bug
The latest nightly version does not support `typing.List` in combination with `from __future__ import annotations`, but requires the use of `list`. While I agree that using `list` is better and more modern, this introduces a breaking change that makes it difficult to keep the custom ops working both on older and newer pytorch versions.
This change was introduced 5 days ago with this commit: https://github.com/pytorch/pytorch/commit/a79100ab11e09473d7456bc02de36de629d9db62#diff-09dd038b7467c407d31c7feb905906037d68d4a419a714bb1f56cee93976de39
My problem with that is: Using List[T] as annotation now gives me the following error:
```
Traceback (most recent call last):
File "/weka/mathias/kuma/.venv/lib/python3.12/site-packages/torch/_library/infer_schema.py", line 65, in convert_type_string
return eval(annotation_type)
^^^^^^^^^^^^^^^^^^^^^
File "<string>", line 1, in <module>
NameError: name 'List' is not defined. Did you mean: 'list'?
```
I could simply change all annotations for `list`, but then it doesn't work anymore with older pytorch versions, requring to create a complete clone of the custom op module.
I already tried using typing Aliases: use `type List = list if version... else type List = typing.List`, but that doesn't work either, because the types are evaluated as a raw string - using capital L in the beginning doesn't work.
**Turns out that typing.List only fails when importing future annotations.**
So, not sure if you want to fix, but since it did work before and would only require reverting one line of code, I wanted to open this issue to discuss if you want to support this case or not.
My workaround now was to remove the `from __future__ import annotations` from these files and change some types.
The line that need to be changed
https://github.com/pytorch/pytorch/blob/a79100ab11e09473d7456bc02de36de629d9db62/torch/_library/infer_schema.py#L6
to the state before this commit:
https://github.com/pytorch/pytorch/commit/a79100ab11e09473d7456bc02de36de629d9db62#diff-09dd038b7467c407d31c7feb905906037d68d4a419a714bb1f56cee93976de39
Use this script to reproduce the error:
```
from __future__ import annotations
import torch
from typing import List
@torch.library.custom_op("test::custom_op_list", mutates_args=())
def my_custom_op(
x: torch.Tensor,
) -> List[torch.Tensor]:
return [torch.randn_like(x)]
```
### Versions
Pytorch 2.7.0.dev20250122+cu124
Python 3.12.6
Ubuntu 20.04
(collect_env.py crashes when using uv)
cc @ezyang @gchanan @zou3519 @kadeng @msaroufim @chauhang @penguinwu @bdhirsh @yf225 | high priority,triaged,module: custom-operators,oncall: pt2,module: pt2-dispatcher | low | Critical |
2,806,850,843 | vscode | We were unable to automatically build your code. Please replace the call to the autobuild action with your custom build steps. We were unable to automatically build your code. Please change the build mode for this language to manual and specify build steps for your project. See https://docs.github.com/en/code-security/code-scanning/troubleshooting-code-scanning/automatic-build-failed for more information. Encountered a fatal error while running "/opt/hostedtoolcache/CodeQL/2.20.1/x64/codeql/codeql database trace-command --use-build-mode --working-dir /home/runner/work/todo/todo -O=cpp.trap.cache.dir=/home/runner/work/_temp/trapCaches/cpp -O=cpp.trap.cache.bound=1024 -O=cpp.trap.cache.write=true /home/runner/work/_temp/codeql_databases/cpp". Exit code was 2 and error was: A fatal error occurred: Exit status 1 from command: [/opt/hostedtoolcache/CodeQL/2.20.1/x64/codeql/cpp/tools/autobuild.sh]. See the logs for more details. | https://github.com/latamexiovas10/todo/actions/runs/12929318719/job/36058459946#step:7:161 | info-needed,triage-needed | low | Critical |
2,806,857,981 | react | CSP violation | As per guidelines from security team, framework team wants to enable Content Security Policy (CSP ) at the gateway layer to enforce good web-security practice. In 2412, framework team evaluated the impact of enabling lenient CSP at the gateway layer. As part of this exercise, we enabled CSP in the gateway layer in Report only mode and collected all CSP violations. we found that
node_modules/react-dom/cjs/react-dom-server-legacy.browser.development.js
node_modules/react-dom/cjs/react-dom-server-legacy.browser.production.min.js
node_modules/react-dom/cjs/react-dom-server-legacy.node.development.js
node_modules/react-dom/cjs/react-dom-server.browser.development.js
node_modules/react-dom/cjs/react-dom-server.node.development.js
node_modules/react-dom/umd/react-dom-server-legacy.browser.development.js
node_modules/react-dom/umd/react-dom-server-legacy.browser.production.min.js
node_modules/react-dom/umd/react-dom-server.browser.development.js
is using eval/dynamic function which is flagged as CSP violation. To fix this, we need to replace usage of eval/dynamic function with something safe. | Status: Unconfirmed | medium | Minor |
2,806,867,847 | vscode | Allow multiple lines per StickyScroll entry | <!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
Some coding conventions mandate opening/closing braces to be on a separate line. For example here's a piece of `PostgreSQL` code:
<details>
<summary>Wall of C code</summary>
```C
RelFileNumber
GetNewRelFileNumber(Oid reltablespace, Relation pg_class, char relpersistence)
{
RelFileLocatorBackend rlocator;
char *rpath;
bool collides;
ProcNumber procNumber;
/*
* If we ever get here during pg_upgrade, there's something wrong; all
* relfilenumber assignments during a binary-upgrade run should be
* determined by commands in the dump script.
*/
Assert(!IsBinaryUpgrade);
switch (relpersistence)
{
case RELPERSISTENCE_TEMP:
procNumber = ProcNumberForTempRelations();
break;
case RELPERSISTENCE_UNLOGGED:
case RELPERSISTENCE_PERMANENT:
procNumber = INVALID_PROC_NUMBER;
break;
default:
elog(ERROR, "invalid relpersistence: %c", relpersistence);
return InvalidRelFileNumber; /* placate compiler */
}
....
}
```
</details>
The StickyScroll makes reading this kind of code really nice, but sometimes there are extremely long `if` statements and having those Stick is really nice. However the `outlineModel` doesn't really care about the `if` statements. The next best alternative would be the `foldingProviderModel`, but that doesn't work well with this coding convention. The Sticky looks like this:
```C
{
{
{
/* some code here */
...
```
And surely if one were to fold the opening braces to the same line, it'd look way more useful:
```C
GetNewRelFileNumber(Oid reltablespace, Relation pg_class, char relpersistence){
switch (relpersistence){
/* you are here */
```
## Feature:
Allow including "lines before" and "lines after" into the sticky scroll, or a hack to "collapse braces" or something, although that is too language specific, unfortunately. | feature-request,editor-sticky-scroll | low | Critical |
2,806,891,699 | react | [React 19] Without extra Generic Type to useActionState hook | ## Summary
Hi everyone!
I have a question about the types from the new hook 'useActionState'.
Actually, I'm working on a project and I need some extra types for my action function

Why don't we have an extra Generic Type to return from the action? Does the return really need to be equal to the initial state?
In a real use case, the return from the action can be different from the entry data, such as an error object like:
```js
// INPUT DATA
{
email: '[email protected]'
}
// OUTPUT DATA
{
email: {
value: '[email protected]'
state: {
error: {
message: null
}
}
}
}
```
**Reference**
- https://react.dev/reference/react/useActionState
- https://react.dev/blog/2024/12/05/react-19#whats-new-in-react-19
<!--
Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a
repository on GitHub, or provide a minimal code example that reproduces the
problem. You may provide a screenshot of the application if you think it is
relevant to your bug report. Here are some tips for providing a minimal
example: https://stackoverflow.com/help/mcve.
-->
| React 19 | medium | Critical |
2,806,902,523 | PowerToys | Last Modifed Date Is Sometimes More Appropriate Than File Creation Date | ### Description of the new feature / enhancement
I shoot many videos with my camera. Then I remove the memory card, insert into a reader on my Windows computer, and use the Sony utility to copy them to a Windows volume. The copy modifies the file creation date.
If I let videos accumulate for a few days before copying them to Windows, they all end up with the same creation date. However, the last modified date remains correct, which is the date each video was made.
### Scenario when this would be used?
It might be appropriate to use file creation date in some circumstances, but anyone who is copying media from a memory card to their computer will want to use last modified date instead. This should be an option.
### Supporting information
* https://www.mslinn.com/git/1600-git-ls-files.html#ucff
* https://support.d-imaging.sony.co.jp/www/disoft/int/download/playmemories-home/win/en/index.html | Needs-Triage | low | Minor |
2,806,914,239 | next.js | pnpm workspace package with non shared lockfile break dev mode | ### Link to the code that reproduces this issue
https://github.com/LukasGerm/pnpm-next-test
### To Reproduce
1. pnpm i
2. Go into the packages/test-package and run pnpm build
3. Go to apps/my-app and run pnpm dev
### Current vs. Expected behavior
Observe that
<img width="397" alt="Image" src="https://github.com/user-attachments/assets/e4808a47-5c0e-4e65-b1e1-e26bf882878f" />
Expected would be that this works. Building the application works btw.
### Provide environment information
```bash
Operating System:
Platform: darwin
Arch: arm64
Version: Darwin Kernel Version 24.1.0: Thu Oct 10 21:06:57 PDT 2024; root:xnu-11215.41.3~3/RELEASE_ARM64_T6041
Available memory (MB): 24576
Available CPU cores: 12
Binaries:
Node: 20.9.0
npm: 10.1.0
Yarn: N/A
pnpm: 9.15.4
Relevant Packages:
next: 15.1.6 // Latest available version is detected (15.1.6).
eslint-config-next: 15.1.6
react: 19.0.0
react-dom: 19.0.0
typescript: 5.7.3
Next.js Config:
output: N/A
```
### Which area(s) are affected? (Select all that apply)
Turbopack
### Which stage(s) are affected? (Select all that apply)
next dev (local)
### Additional context
I also tested without turbopack, but the same issue | Turbopack,linear: turbopack,Package Managers | low | Minor |
2,806,923,162 | flutter | FlutterEngineGroup isolates run in same thread but should not | ### Steps to reproduce
I am utilizing FlutterEngineGroup between my main and my worker isolates on mobile devices so that I can share [complex objects](https://github.com/flutter/flutter/issues/152484).
I am also using a fork of flutter_isolate plugin, so that I can have callbacks from native side to the respective isolate.
With their recent change, it looks like all the isolates which are spawned from the same FlutterEngineGroup, run in the same thread. I verified this by calling `syscall(SYS_gettid)` via FFI and it came back with same numbers. More context [here](https://github.com/rmawatson/flutter_isolate/issues/160).
Is this expected? Do I need to spawn a separate native thread in order to run the new engine (within same engine group) in a separate thread?
### Expected results
Isolates within same FlutterEngineGroup should run in separate threads
### Actual results
Isolates within same FlutterEngineGroup run in same thread.
### Code sample
<details open><summary>Code sample for retrieving thread id</summary>
```c++
#include <unistd.h>
#include <sys/syscall.h>
#include <stdint.h>
int64_t get_thread_id() {
return (int64_t)syscall(SYS_gettid);
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
[Upload media here]
</details>
### Logs
<details open><summary>Logs</summary>
```console
[Paste your logs here]
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[Paste your output here]
```
</details>
| waiting for customer response,in triage | low | Major |
2,806,939,325 | flutter | [file_selector] Add option to customize the directory selection dialog title | ### Use case
When calling `getDirectoryPath` I couldn't find a way to customize the dialog title.
This would be important to give context to the user about the purpose of the directory selection.
### Proposal
A ppssible solution is to add an optional parameter to the method:
`getDirectoryPath({String? dialogTitle})` | package,team-ecosystem,P3,p: file_selector,triaged-ecosystem | low | Minor |
2,806,954,203 | godot | Unrecognized UID: "uid://a" | ### Tested versions
4.4 beta1
### System information
W10
### Issue description
When opening project
```
Unrecognized UID: "uid://a"
```
is printed to the console.
I tracked the source to GDScript parser, but no idea what causes it.
### Steps to reproduce
Open MRP.
### Minimal reproduction project (MRP)
[Godot-Instance-Dock.zip](https://github.com/user-attachments/files/18521292/Godot-Instance-Dock.zip)
Includes .godot folder, because the error is rather finicky. There is a chance you still won't be able to reproduce it. | bug,topic:gdscript,needs testing,regression | low | Critical |
2,806,971,341 | ui | [bug]: When trying to initialize `shadcn-ui` in an Astro project, I encounter an error: `ENOENT: no such file or directory`, related to the creation of the folder `src/*/lib`. | ### Describe the bug
### Description
When trying to initialize `shadcn-ui` in an Astro project, I encounter an error: `ENOENT: no such file or directory`, related to the creation of the folder `src/*/lib`.
> ENOENT: no such file or directory, mkdir 'C:_beppe\Astro_dev\ahorros-familiares-astro-sphere\src*\lib' of my project
I already have a project created based on the AstroTheme called 'Astro-sphere'. I'm not sure if this could be causing the issue when installing shadcn on top of it.
I have tried creating the folders manually, but it still doesn't work. What surprises me is that it mentions a path like src/*/lib, but I don't have this structure in my Astro project. I'm not sure what the * is supposed to refer to
### Affected component/components
Installation on Astro
### How to reproduce
### Steps to reproduce
1. Create an Astro project using `npm create astro@latest`.
2. Configure Tailwind CSS according to the documentation.
3. Run the command `npx shadcn@latest init`.
4. Choose the default options and force dependency installation (`--force`).
5. The following error occurs:
### Codesandbox/StackBlitz link
_No response_
### Logs
```bash
PS C:\_beppe\Astro\_dev\ahorros-familiares-astro-sphere> npx shadcn@latest init
✔ Preflight checks.
✔ Verifying framework. Found Astro.
✔ Validating Tailwind CSS.
✔ Validating import alias.
√ Which style would you like to use? » New York
√ Which color would you like to use as the base color? » Neutral
√ Would you like to use CSS variables for theming? ... no / yes
✔ Writing components.json.
✔ Checking registry.
✔ Updating tailwind.config.mjs
✔ Updating src\styles\global.css
Installing dependencies.
It looks like you are using React 19.
Some packages may fail to install due to peer dependency issues in npm (see https://ui.shadcn.com/react-19).
√ How would you like to proceed? » Use --force
✔ Installing dependencies.
⠹ Updating files.
Something went wrong. Please check the error below for more details.
If the problem persists, please open an issue on GitHub.
ENOENT: no such file or directory, mkdir 'C:\_beppe\Astro\_dev\ahorros-familiares-astro-sphere\src\*\lib'
```
### System Info
```bash
### Additional details
- **Operating system**:Windows 10
- **Node.js version**: v20.11.0
- **shadcn-ui version**: Latest (`npx shadcn@latest`)
- **Astro version**: "astro": "^4.15.12",
- **Tailwind CSS version**: "^3.4.13"
```
### Before submitting
- [x] I've made research efforts and searched the documentation
- [x] I've searched for existing issues | bug | low | Critical |
2,806,976,716 | electron | document.fragmentDirective has no properties inside webview | ### Preflight Checklist
- [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project.
- [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [x] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success.
### Electron Version
34.0.1
### What operating system(s) are you using?
Ubuntu
### Operating System Version
23.04
### What arch are you using?
x64
### Last Known Working Electron version
_No response_
### Expected Behavior
When calling `document.fragmentDirective`, it should have an `items` and a `createSelectorDirective` property[1]. This is the case when calling it from a browserWindow's console or webcontentsview,
[1] details https://github.com/WICG/scroll-to-text-fragment/blob/main/fragment-directive-api.md
### Actual Behavior
When calling it from a webview's console then `document.fragmentDirective` has no properties, which mean you can't create text fragment links from pages inside webviews
### Testcase Gist URL
_No response_
### Additional Information
This is a relatively new API. | component/webview,bug :beetle:,platform/all,has-repro-gist,34-x-y,35-x-y | low | Critical |
2,806,989,174 | terminal | User32.dll Import Cause Windows Terminal Crash | ### Windows Terminal version
1.21.3231.0
### Windows build number
10.0.26100.0
### Other Software
_No response_
### Steps to reproduce
## Problem Statement
It seems like the use of user32.dll could cause stability issue on Windows terminal. Eventually, it will cause the crash.
This issue only happens with a small possibility when the command is executed (<0.1% chance/cycle). But when we put this in a while loop a run the same script across 200+ machines, it will happen everyday.
In this case, I isolated the code to reproduce this problem.
* run following code as administrator.
* start another 15 terminals to run that in parallel.
## Code to reproduce
```powershell
function Out-ProcessEx {
param
(
[Parameter(Mandatory, ValueFromPipeline)]
$process
)
process {
#Get Window Title, We use this function because default PowerShell function cannot get
#title for hidden window
$wTitle = Get-WindowTitleFromProcessId -processId $process.Id
# update MainWindowTitle
$field = $process.GetType().GetField("mainWindowTitle", [System.Reflection.BindingFlags]::Instance -bor [System.Reflection.BindingFlags]::NonPublic -bor [System.Reflection.BindingFlags]::Public)
$field.SetValue($process, $wTitle)
$process
}
}
function Get-WindowHandleFromProcessId {
param (
[Parameter(Mandatory = $true)]
[int]$ProcessId
)
# Define the necessary functions from user32.dll
Add-Type @"
using System;
using System.Runtime.InteropServices;
public class User32 {
public delegate bool EnumWindowsProc(IntPtr hWnd, IntPtr lParam);
[DllImport("user32.dll", SetLastError = true)]
[return: MarshalAs(UnmanagedType.Bool)]
public static extern bool EnumWindows(EnumWindowsProc lpEnumFunc, IntPtr lParam);
[DllImport("user32.dll", SetLastError = true)]
public static extern int GetWindowThreadProcessId(IntPtr hWnd, out int lpdwProcessId);
}
"@ -ErrorAction SilentlyContinue
$Script:windowHandle = 0
$callback = [User32+EnumWindowsProc] {
param ($hWnd, $lParam)
$processId = 0
[User32]::GetWindowThreadProcessId($hWnd, [ref]$processId) | Out-Null
if ([int]($processId) -eq [int]($lParam)) {
$Script:windowHandle = $hWnd
return $false # Stop enumeration
}
return $true # Continue enumeration
}
[User32]::EnumWindows($callback, [IntPtr]$ProcessId) | Out-Null
return $Script:windowHandle
}
function Get-WindowTitleFromProcessId {
param (
[Parameter(Mandatory = $true)]
[Int]$processId
)
# Define the necessary functions from user32.dll
Add-Type @"
using System;
using System.Text;
using System.Runtime.InteropServices;
public class WindowTitleFromHandle {
[DllImport("user32.dll", CharSet = CharSet.Auto, SetLastError = true)]
public static extern int GetWindowText(IntPtr hWnd, StringBuilder lpString, int nMaxCount);
[DllImport("user32.dll", SetLastError = true)]
public static extern int GetWindowTextLength(IntPtr hWnd);
}
"@ -ErrorAction SilentlyContinue
$WindowHandle = Get-WindowHandleFromProcessId -ProcessId $processId
$length = [WindowTitleFromHandle]::GetWindowTextLength($WindowHandle)
if ($length -gt 0) {
$builder = New-Object System.Text.StringBuilder -ArgumentList ($length + 1)
[WindowTitleFromHandle]::GetWindowText($WindowHandle, $builder, $builder.Capacity) | Out-Null
return $builder.ToString()
}
else {
return "No title found"
}
}
while($true)
{
$s1 = Get-Process|Out-ProcessEx
#start 16 process and see......
}
```
## Error Log


### Expected Behavior
16 instances of windows terminal should persist after 20 minutes
### Actual Behavior
* The crash issue typically occur after ~ 15-20mins. | Issue-Bug,Needs-Triage | low | Critical |
2,806,995,030 | terminal | Can't disable startup | ### Windows Terminal version
1.21.3231.0
### Windows build number
10.0.26100.0
### Other Software
Docker (disabled on startup, does not launch)
AthenaOS WSL
### Steps to reproduce
1. Install WSL (any distro, currently using AthenaOS WSL)
2. Install Windows Terminal
3. Install Docker Desktop
4. Disable docker on startup
5. Disable Launch on Machine startup in Windows Terminal settings

7. Check `shell:startup`

8. Check `shell:Common Startup`

9. Check task manager startup apps

10. Check Windows Settings App -> Apps -> Startup (NOTICE HERE, Terminal is checked on and greyed out so it can't be disabled??)

11. On Fresh windows start, terminal auto opens and launches WSL
Things I have tried to alleviate:
1. I have disabled WSL, which still causes Terminal to launch at startup, but will just launch with a WSL error because WSL is not running
2. I have turned off showing recent files in explorer
3. I have ensured Linux subsystem folders are not expanded or manually pinned to quick start
4. Ensured docker desktop with WSL support does not run at startup
5. https://www.winhelponline.com/blog/wp-content/uploads/2023/06/store_appx_startup_fix.zip registry hack which allows me to actually change Startup Apps under Settings App to disable Windows Terminal
### Expected Behavior
I expect Windows Terminal to not autostart when computer starts, I also expect to be able to disable Windows Terminal from Settings App -> Apps -> Startup, currently it is enabled and greyed out with no way to disable it.
### Actual Behavior
Windows terminal launches at startup everytime. | Issue-Bug,Needs-Triage | low | Critical |
2,807,025,153 | rust | Tracking issue for release notes of #126604: Uplift `clippy::double_neg` lint as `double_negations` |
This issue tracks the release notes text for #126604.
### Steps
- [ ] Proposed text is drafted by PR author (or team) making the noteworthy change.
- [ ] Issue is nominated for release team review of clarity for wider audience.
- [ ] Release team includes text in release notes/blog posts.
### Release notes text
The responsible team for the underlying change should edit this section to replace the automatically generated link with a succinct description of what changed, drawing upon text proposed by the author (either in discussion or through direct editing).
````markdown
# Category (e.g. Language, Compiler, Libraries, Compatibility notes, ...)
- [Uplift `clippy::double_neg` lint as `double_negations`](https://github.com/rust-lang/rust/pull/126604)
````
> [!TIP]
> Use the [previous releases](https://doc.rust-lang.org/nightly/releases.html) categories to help choose which one(s) to use.
> The category will be de-duplicated with all the other ones by the release team.
>
> *More than one section can be included if needed.*
### Release blog section
If the change is notable enough for inclusion in the blog post, the responsible team should add content to this section.
*Otherwise leave it empty.*
````markdown
````
cc @kadiwa4, @nnethercote -- origin issue/PR authors and assignees for starting to draft text
| T-lang,relnotes,needs-triage,relnotes-tracking-issue | low | Minor |
2,807,038,420 | iptv | Edit: Workpoint 23 (1080p) | ### Stream URL (required)
http://appdootv2.dootvde.com:1935/live/50016_workpoint_tv.stream.smil/playlist.m3u8
### Channel ID
Workpoint23HD.th
### Channel Name
Workpoint 23
### Quality
1080p
### Label
None
### Timeshift
_No response_
### HTTP User Agent
_No response_
### HTTP Referrer
_No response_
### Notes
_No response_
### Contributing Guide
- [x] I have read [Contributing Guide](https://github.com/iptv-org/iptv/blob/master/CONTRIBUTING.md) | streams:edit | low | Minor |
2,807,051,525 | vscode | Provide way to disable Copilot Free signup prompt for enterprises | <!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
In many enterprises the free version of GitHub Copilot is disallowed for data security reasons. Therefore we do not want users to be prompted to sign up to Copilot Free when they start VSCode.
We cannot work around this by removing the Copilot extension as we do allow Copilot enterprise, just not the free version. | workbench-copilot | low | Minor |
Subsets and Splits