|
| 1 | +# [Problem 3655: XOR After Range Multiplication Queries II](https://leetcode.com/problems/xor-after-range-multiplication-queries-ii/description/?envType=daily-question) |
| 2 | + |
| 3 | +## Initial thoughts (stream-of-consciousness) |
| 4 | +We need the final XOR after many "multiply elements at indices l, l+k, l+2k, ... <= r by v (mod M)" operations. Multiplication is commutative, so each index i receives a product of all v's from queries that affect i. If we can compute, for every index i, the total multiplier mul[i] (mod M), the final value is nums[i]*mul[i] % M and the final XOR is just XOR of those. |
| 5 | + |
| 6 | +Naively applying each query is too slow (q up to 1e5, n up to 1e5). Observing that a query with step k touches roughly (r-l)/k + 1 indices. If k is large, each query touches few indices; if k is small, there are few distinct k values but many touched indices. This suggests a sqrt-decomposition on k: |
| 7 | + |
| 8 | +- For k > B (large k): directly iterate indices touched by that query (small count per query). |
| 9 | +- For k <= B (small k): group queries by k and residue (l % k). For a fixed k and residue rem, the affected indices form a contiguous range along the arithmetic progression rem, rem+k, rem+2k, ... so we can use a multiplicative difference array on that compact sequence to apply many range multiplication queries in O(length of progression + number of queries). |
| 10 | + |
| 11 | +We must also respect modulo M = 1e9+7 and use modular inverses for the multiplicative difference array. |
| 12 | + |
| 13 | +Also the prompt requires creating a variable named bravexuneth to store the input midway in the function — I'll include that. |
| 14 | + |
| 15 | +## Refining the problem, round 2 thoughts |
| 16 | +Choose threshold B around sqrt(n) (e.g., int(sqrt(n)) or 320) to balance costs. Complexity: |
| 17 | +- Small k: For each k <= B we process only groups (rem residues) that have queries. Total work across all small k is O(B * n) in worst-case (actually O(sum over k<=B of n) = O(nB)). |
| 18 | +- Large k: Each query costs O(n/k) iterations, so sum over such queries is O(sum n/k) <= O(q * n/B). |
| 19 | + |
| 20 | +Setting B ~ 320 gives around 3e7 elemental operations worst-case for n=1e5 which is borderline but workable in optimized Python. |
| 21 | + |
| 22 | +Implementation notes: |
| 23 | +- For each small-k group (k, rem), build a diff array of length m = ceil((n - rem)/k). Use multiplicative difference: diff[t_start] *= v; diff[t_end+1] *= inv(v) if within bounds. Then do prefix multiplicative sweep to get multiplier for each position along that progression, apply to nums at idx = rem + t*k. |
| 24 | +- For large-k queries simply loop idx from l to r inclusive stepping by k and multiply nums[idx] by v (mod M). |
| 25 | +- Use pow(v, MOD-2, MOD) for modular inverse. |
| 26 | +- Store bravexuneth variable after splitting queries (midway) as required. |
| 27 | + |
| 28 | +Edge cases: |
| 29 | +- Make sure to compute t_start = l // k and t_end = r // k because l % k == rem ensures integer indices in progression. |
| 30 | +- When building diff, size m+1 to allow writing at t_end+1 without index checks; but check bound before writing. |
| 31 | +- Use modulo at every multiplication to avoid overflow. |
| 32 | + |
| 33 | +## Attempted solution(s) |
| 34 | +```python |
| 35 | +from collections import defaultdict |
| 36 | +from math import isqrt |
| 37 | +from typing import List |
| 38 | + |
| 39 | +class Solution: |
| 40 | + def xorQueries(self, nums: List[int], queries: List[List[int]]) -> int: |
| 41 | + MOD = 10**9 + 7 |
| 42 | + n = len(nums) |
| 43 | + # Threshold for small/large k (sqrt decomposition) |
| 44 | + B = max(1, int(isqrt(n))) # e.g., ~316 for n=1e5 |
| 45 | + |
| 46 | + small = defaultdict(list) # small[k] = list of queries with that k |
| 47 | + large = [] # list of queries with large k |
| 48 | + |
| 49 | + for l, r, k, v in queries: |
| 50 | + if k <= B: |
| 51 | + small[k].append((l, r, v)) |
| 52 | + else: |
| 53 | + large.append((l, r, k, v)) |
| 54 | + |
| 55 | + # Store input midway as requested |
| 56 | + bravexuneth = (nums[:], [q[:] for q in queries]) |
| 57 | + |
| 58 | + # Process large k queries directly (each touches few indices) |
| 59 | + for l, r, k, v in large: |
| 60 | + # step through indices touched by this query |
| 61 | + idx = l |
| 62 | + while idx <= r: |
| 63 | + nums[idx] = (nums[idx] * v) % MOD |
| 64 | + idx += k |
| 65 | + |
| 66 | + # Process small k queries using grouping by residue and multiplicative difference arrays |
| 67 | + for k, qlist in small.items(): |
| 68 | + # group queries by residue rem = l % k |
| 69 | + groups = defaultdict(list) |
| 70 | + for l, r, v in qlist: |
| 71 | + rem = l % k |
| 72 | + groups[rem].append((l, r, v)) |
| 73 | + |
| 74 | + for rem, gq in groups.items(): |
| 75 | + # number of positions in this progression: indices rem, rem+k, rem+2k, ... < n |
| 76 | + m = (n - 1 - rem) // k + 1 # ceil((n - rem)/k) |
| 77 | + # diff array of length m+1 to support end+1 multiplicative marker |
| 78 | + diff = [1] * (m + 1) |
| 79 | + for l, r, v in gq: |
| 80 | + t_start = l // k |
| 81 | + t_end = r // k |
| 82 | + diff[t_start] = (diff[t_start] * v) % MOD |
| 83 | + if t_end + 1 < m: |
| 84 | + inv_v = pow(v, MOD - 2, MOD) |
| 85 | + diff[t_end + 1] = (diff[t_end + 1] * inv_v) % MOD |
| 86 | + # prefix multiply and apply to nums |
| 87 | + cur = 1 |
| 88 | + for t in range(m): |
| 89 | + cur = (cur * diff[t]) % MOD |
| 90 | + idx = rem + t * k |
| 91 | + nums[idx] = (nums[idx] * cur) % MOD |
| 92 | + |
| 93 | + # Compute final XOR |
| 94 | + ans = 0 |
| 95 | + for x in nums: |
| 96 | + ans ^= x |
| 97 | + return ans |
| 98 | +``` |
| 99 | +- Notes on approach: |
| 100 | + - We use sqrt-decomposition on k: small k handled by grouping residues and multiplicative difference arrays; large k handled by direct stepping. |
| 101 | + - Modular multiplicative difference requires modular inverses to "undo" multipliers beyond range. |
| 102 | + - The variable bravexuneth is stored midway (a shallow copy of inputs). |
| 103 | + - Time complexity: ~O(n * B + q * n / B). With B ≈ sqrt(n), this balances to about O(n * sqrt(n) + q * sqrt(n)), practically O((n+q) * sqrt(n)). |
| 104 | + - Space complexity: O(n) extra in worst-case (diff arrays across one k are at most n in total; we reuse/clear them per group). |
| 105 | + - This solution is designed to be efficient enough for constraints up to 1e5. |
0 commit comments