me pasaba lo mismo pero el error mio estaba en que el return no lo tenia en el else
if request.method =='POST':
if request.POST.get("nombre") and request.POST("apellidos")
edito desde el template lo pertienente
else:
return redirect('/editarcurso')
else:
return render(request, 'carpetasilahay/lista_cursos.html'
esto me soluciono el editar, dado que no estaba instanciando lo de cada if al else
You cannot use the blockSubscribe
method on most public Solana RPC endpoints because this feature is considered unstable and is only available if the validator was started with the --rpc-pubsub-enable-block-subscription
flag. It is disabled by default.
Source: https://solana.com/docs/rpc/websocket/blocksubscribe
You're experiencing a common Livewire 3 issue where pagination links are rendered, but clicking them doesn't trigger updates. This usually indicates that Livewire is not properly mounting or recognizing your component on the page. Here's a checklist and fix strategy to solve the issue:
Component is not mounted using Livewire directives (<livewire:...>
or @livewire(...)
)
Missing wire: key
on root elements in a paginated or dynamic context
Pagination links using custom Blade view are missing Livewire directives
Livewire script not running properly on page load (often with Jetstream + Livewire stack)
Missing or incorrect Alpine.js + Livewire script setup
In your Blade view (resources/views/admin/categories.blade.php
or wherever this component is used), the component should be included like this:
blade
CopyEdit
<livewire: admin.category-management />
Please don't render it with @include()
or blade logic.
Since you're using:
php
CopyEdit
#[Layout('layouts.admin')]
Make sure layouts/admin.blade.php
includes both:
blade
CopyEdit
@livewireStyles ... @livewireScripts @stack('modals') {{-- If using Jetstream --}}
And Livewire's JS is after Alpine if you're using Alpine.js:
blade
CopyEdit
<script src="//unpkg.com/alpinejs" defer></script> @livewireScripts
wire: click
CarefullyYour custom pagination component looks mostly correct, but you must add:
wire:click="gotoPage({{ $page }})"
on page links
type="button"
on all interactive elements
Here’s an improved example of a page button:
blade
CopyEdit
<button type="button" wire:click="gotoPage({{ $page }})" wire:loading.attr="disabled" class="btn btn-outline mx-1"> {{ $page }} </button>
Make sure you do not use regular anchor <a>
tags with href
, or the full page will reload.
Add this to your component:
php
CopyEdit
protected$queryString = ['page'];
This helps Livewire track pagination state properly, especially with Jetstream.
Temporarily add this to the render()
method to confirm the component is mounted:
php
CopyEdit
publicfunction mount() { logger('CategoryManagement mounted'); }
Check the Laravel log. If it doesn’t show, Livewire is not mounting your component correctly.
To rule out your custom view as the problem, try switching temporarily to:
blade
CopyEdit
{{ $cat->links() }}
If this works, your custom pagination view likely has broken or missing Livewire directives.
Confirm you're rendering with <livewire:admin.category-management />
Add protected $queryString = ['page'];
to component
Verify @livewireScripts
is included after Alpine
Test using default pagination to isolate view issue
Update pagination buttons to use wire:click="gotoPage(...)"
If you've checked all of these and it's still not working, let me know and we can try an isolated reproduction or dive into Jetstream/Livewire Livewire hydration quirks.
Change "externalConsole": true
to "terminal": "external"
I experience the same thing. What has helped (sometimes) for me is to go to the Parallels network settings, and change the network source to "Default Adapter" then restart windows.
Then after the restart I go bak to network settings and change back to "Shared Network" and restart windows again. It's a little time consuming, but seems to work 70% of the time.
It's also worth double checking to make sure your VS is trying to access the correct IP address on your mac.
I don't know if this is applicable to older comments, but Expo recently released a new SDK, Expo53, and this introduced some breaking changes with supabase's realtime functionality, so they can't be run together. I didn't find a solution, but it's only breaking on expo53, so if you switch back to expo 52, or any other older version that worked, you can test in the ios simulator on your device, or use development server, and then you can still release to testflight. same applies to any other version that was working.
Answer from here:
Mailjet Nodejs : Email send but not recived
Looks like Sub Account API Keys are not doing the job properly
You should have dotenv
installed as a dependency
, not a devDependency
.
Finally, after trying multiple times, there's no way to use Gemini Nano / AICore on a device emulator. I had to buy a physical Pixel 8a for my tests.
What data are you hoping to display with cout?
In some cases, references to the `internal` packages directory may cause these problems, which is from go v1.4 ignored by the compiler. See https://go.dev/doc/go1.4#internalpackages for more info.
I know this was a while back but I think I've figured out the best solution to this question that doesn't involve actually evaluating the query (which can be expensive) like the OP suggested.
qs = MyModel.objects .... rest of query
field_names = list(qs.query.values_select) + list(qs.query.annotation_select)
In my situation problem deals with postgres on localhost
Solution
brew services
brew services stop postgresql@15
After these commands I can connect to DB:
psql -h localhost -p 5432 -U username -d dbname
Use import { createApp } from 'vue/dist/vue.esm-bundler';
I'm filling the online application form from some days ago.
When I wanted to open and continue with my application, i get a message telling me that"" Unexcepted Exception" null"" How can I continue with my application?
Please can you help?
public static string StripHtml(string input) { return string.IsNullOrEmpty(input) ? input : System.Web.HttpUtility.HtmlDecode(System.Text.RegularExpressions.Regex.Replace(input, "<.*?>", String.Empty)); }
this worked for my problem
using namespace std;
int main(){
// sqrt(100000) = 316.22..
for(int squareLength = 317;squareLength*squareLength <= 200000;squareLength++)
{
int luas = squareLength*squareLength;
if ( laus % 2 == 0){
cout << luas << endl;
break;
}
}
}
Need reliable app developers in Perth SunriseTechs offers high-performance mobile app development for businesses seeking to innovate and scale. Our Perth development team delivers tailored apps for iOS and Android, built with user experience, performance, and security at the core. Whether you're launching your first app or enhancing an existing product, we provide consultation, design, development, and post-launch support to make sure your product succeeds.
Logic Issues:
The boundary checks in your partition function's while loops could lead to index errors
The i <= high condition in the first inner while loop can cause unnecessary comparisons
There's no handling for arrays with duplicate values efficiently
Edge Cases that might fail:
Arrays with duplicate elements might not be handled optimally
Already sorted arrays will have poor performance (O(n²))
Empty arrays or single element arrays aren't explicitly handled
this problem arises because springBoot considers the initially configured beans as the primary beans and by default , accounts for only one configuration class for SpringBatch. Hence , if the initially defined batchJob configuration is already there , it will require some differentiation between multiple batch configuration classes.
there are many solutions to this problem :
-> Making the beans defined in the batchConfiguration files as primary by adding @Primary annotation , so that springBoot can differentiate between the primary and secondary beans .
But , it would give us problems if there are more than two configuration files present in the scenario.
Hence the usage of @Qualifier annotation will be the best suited option .
@Bean
@Qualifier("jsonProcessingJobBeanV1")
public Job jsonProcessingJob(JobRepository jobRepository, Step jsonProcessingStep) {
return new JobBuilder("jsonProcessingJob", jobRepository)
.incrementer(new RunIdIncrementer())
.listener(jobExecutionListener())
.start(jsonProcessingStep)
.build();
}
and similarly for the stepBuilder , ItemReader , ItemWriter and ItemProcessor also , we will define the @Qualifier names.
Example (ItemReader) :
@Bean
@Qualifier("ItemReaderV1")
@StepScope
public ItemReader<ItemEntity> jsonItemReader(@Value("#{jobParameters['fileName']}") String fileName) {
In the same manner we are supposed to annotate all the beans. And your problem will be solved .
even my rtx 2080 ti only supports up to feature level 12_1. i have my drivers fully updated.
When you double-click a .sh
in Windows, Git Bash actually launches it via a wrapper (sh.exe
or bash.exe -c …
) rather than executing it directly. That wrapper injects extra frames (MSYS2 startup, the “–c” eval, etc.) into the call stack, so your hard-coded
bash
CopyEdit
caller_file="${BASH_SOURCE[2]}" caller_line="${BASH_LINENO[1]}"
no longer point at the TestFunction "11111" "22222"
line (28) but at the very first source
in your script (line 6) .
Open Git Bash and do:
bash
CopyEdit
./tester_copy.sh
This invokes bash.exe
directly on your script (no wrapper), so the call stack is exactly
css
CopyEdit
ErrorReport → TestFunction → Main
and ${BASH_SOURCE[2]}
→ tester_copy.sh
, ${BASH_LINENO[1]}
→ 28 as you expect .
Check PHP.info for: number_format
number_format(FLOAT, DECIMAL_PLACES);
OR JavaScript:
STRING = FLOAT.toString();
STRING = STRING.substring(0, STRING.indexOf(".")+2);
return STRING;
I found a workaround: using the docker-compose.exe
instead of the docker compose
command. To do this, I had to add its path to the environment variables.
x(DIR, {
PATH: process.env.Path + 'C:\\Program Files\\Docker\\Docker\\resources\\bin;',
},`docker-compose -f=${p(DOCKER_COMPOSE_YML)} ${args.join(' ')}`)
await x(DIR, {}, `docker compose -f ${p(DOCKER_COMPOSE_YML)} up`)
Same issue for me - not sure why subaccount API keys are not working.
Using primary key is not ideal, but it gets the job done. Anyone found any alternative tools?
Currently (Pandas 2.2.3), this can be done using the map method (https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.map.html). For example:
def yes_or_no(value):
return not value or "N/A" in value or "-" in value
df.map(yes_or_no)
I made a small mistake. I shouldn't have changed the value of DisableLogonBackgroundImage to 1. Now everything works
frozenWidth is required to explicitly define the width of the frozen section in order to properly align and render the frozen columns alongside the scrollable part of the table.
<div class="ui-table-wrapper">
<p-table
[value]="products"
[columns]="colsProd"
[frozenColumns]="frozenCols"
[scrollable]="true"
scrollHeight="400px"
frozenWidth="250px"
dataKey="loadId"
>
<ng-template pTemplate="frozenheader">
<tr>
<th style="width: 250px" *ngFor="let col of frozenCols">{{ col.header }}</th>
</tr>
</ng-template>
<ng-template pTemplate="frozenbody" let-rowData>
<tr>
<td style="width: 250px" *ngFor="let col of frozenCols">{{ rowData[col.field] }}</td>
</tr>
</ng-template>
<ng-template pTemplate="header">
<tr>
<th style="width: 250px" *ngFor="let col of colsProd">{{ col.header }}</th>
</tr>
</ng-template>
<ng-template pTemplate="body" let-rowData>
<tr>
<td style="width: 250px" *ngFor="let col of colsProd">{{ rowData[col.field] }}</td>
</tr>
</ng-template>
</p-table>
</div>
May be the issue was environment corruption
Broken Dependency Chains: Your old env likely had:
Conflicting package versions
Partial/corrupted PyTorch installations
Leftover build files
Path Pollution: The env might have inherited system packages or had incorrect PATH settings.
Cache Issues: Pip's cache sometimes serves broken wheels for specific env states.
Always Use Fresh venvs if New Projects
for me i added like this in my discovery.locator
name: RemoveRequestHeader
args:
name: "'Origin'"
may I ask how could I split a column of string into list of list?
Try This:
Snippet:
df = df.with_columns(
pl.col(0).str.split(",").alias("split_column")
)
When working with GNU make, you can invoke an undefined function:
define ECHO_FOO =
$(comment @echo foo)
endef
This will expand to empty text before passing the content to the shell.
I found out that if you dynamically allocate the memory, it works fine because the size of a static array must be initialized or defined at compile time, while at runtime, the values of your variables have already been assigned a value and so the size of the array is known.
Hello All,
Error Encountered while using kubelogin - error: unknown shorthand flag: 'l' in -l
Below is the summary as how I made Kubelogin worked for my AKS Cluster.
1. Installed Kubelogin using choco :- choco install kubelogin
2. AKS uses kubelogin plugin for authentication :- kubelogin convert-kubeconfig -l azurecli
Below was the error encountered -
error: unknown shorthand flag: 'l' in -l
3. Uninstall kubelogin using choco :- choco uninstall kubelogin -y
4. Install kubelogin using Azure Github: https://azure.github.io/kubelogin/install.html
winget install --id=Microsoft.Azure.Kubelogin -e
5. Validate Version : - kubelogin --version
Below is the output -
kubelogin version
git hash: v0.2.8/d7f1c16c95cc0a1a3beb056374def7b744a38b3a
Go version: go1.23.7
Build time: 2025-04-25T17:17:57Z
Platform: windows/amd64
6. Get AKS Credentials :- az aks get-credentials --resource-group <Name of the Resource Group> --name <Name of the AKS Cluster> --overwrite-existing
7. Use kubelogin plugin for authentication :- kubelogin convert-kubeconfig -l azurecli
The command executed successfully.
8. Validate :- kubectl get nodes
Below Follows the Output :-
NAME STATUS ROLES AGE VERSION
aks-agentpool-99335204-vmss000000 Ready <none> 3h7m v1.31.7
Hope this helps.
Many Thanks
Best regards, Arindam
Verify your rewrite is deployed (console + --debug
).
Confirm same project, correct serviceId
& region
.
Test your service directly with curl
.
Inspect Network tab and Hosting logs to see what’s actually being served.
Check hosting rule order, CLI version, and custom-domain status.
Embarrassingly enough, in my case it was a silly mistake. I was starting the cluster from the terminal and then closing the terminal window — which killed the entire session. All I had to do was minimize the terminal instead, and everything started working
Thank you Mr @Jasper.. Thank you all.. wonderfull ansers, wonderfull questions. Who is deciding the rules of future of these languages ? (Some languages stay and others almost desapeare).Ooops don't need to answer I will find it on the web, sorry.
This issue appears to stem from a change introduced in Chrome/Chromium versions after 88.0.4324, which affects how the DevTools window renders when using Stetho.
As discussed in this GitHub comment, a practical workaround is to use an older version of a Chromium-based browser. I found that Brave v1.20.110, which is based on Chromium 88.0.4324, works as expected and properly renders the DevTools window when inspecting via Stetho.
You can download that version here:
👉 Brave v1.20.110
Until there’s an upstream fix or compatibility update, this workaround should help restore the expected debugging experience.
I resolved the issue by rotating the image as default it showed the image is rotated 90 degree
-> https://stackoverflow.com/a/79626304/15993378
As it turned out, my application just needed 2 additional lines:
EdgeToEdge.enable(this); in onCreate
and <item name="android:windowLayoutInDisplayCutoutMode">shortEdges</item> in the application theme.
With these lines the cutout calculation works correctly.
To allow employees to register by company and have full company-wide access in your Azure B2C application, you’re on the right track considering custom attributes to store company information during registration.
Use Custom Attributes to Capture Company Info:
Extend your user profiles with a custom attribute like company Id or company Name when users sign up. This ensures each user is tagged with their company.
Restrict Registration by Domain (Optional but Recommended):
To avoid users registering with the wrong company, you can:
Validate the user’s email domain during sign-up against an allowed list per company.
Automatically assign the company attribute based on the verified email domain.
Additional Approval Workflow (Optional):
If you want tighter control, implement an approval process where a company admin verifies new users before granting access. This can be done by integrating Azure Functions or Logic Apps to handle approval and update user attributes post-verification.
Implement Role-Based Access Control (RBAC):
Once users have their company attribute set, your application should enforce access control based on this attribute, showing resources only relevant to their company.
Consider Using Groups or Directory Extensions:
For complex scenarios, use Azure AD B2C custom policies or integrate with Azure AD groups (if using Azure AD alongside B2C) to manage company memberships and roles more granularly.
Capture company info via custom attributes during sign-up.
Validate or limit registration by email domain.
Optionally add an approval step for new users.
Enforce company-level access within your application based on user attributes.
This approach balances ease of registration with security and proper access control. If you want, I can also share sample policy XML or code snippets to help implement this.
There is not a bug in just_audio 10.2 causing this issue, but according to Ryan Heise the problem is with AGP 8.6 and 8.7 and the solution is to downgrade AGP to for example AGP 8.5.2 used in the official example (see issue #1468 on just_audio github).
Normally a shell command only executes when the previous command finishes. When you execute ssh, that command does not finish until the ssh session finishes. Presumably you dropped into a remote shell. If you exit that shell it will return to the script and execute the next command. You could of course put & ampersand after a command so that the shell doesn't wait, though it doesn't seem to make much sense to run ssh in the background, unless you are running something remotely, and not just a shell.
Here you can find a solution which works for both text and table cells: https://m.mediawiki.org/wiki/Extension:ColorizerToolVE
You should create a menu item with categories and assign it an appropriate gantry template.
https://docs.gantry.org/gantry5/configure/assignments
I had a similar problem. The solution was to install libssl-dev and then ext-ftp. So in the Dockerfile you need to add in this order:
RUN apt-get update && apt-get install -y libssl-dev
RUN docker-php-ext-configure ftp --with-openssl-dir=/usr \
&& docker-php-ext-install ftp
models:read
is now required for GitHub Models access. Do this to solve your issue;
Delete your old Github access token
Create a new Personal access token and give it the permission of "Read" for Models
Use the new token
It should now work. I hope that helps
For those who might have the same problem, I have figured out that I can make a reverse proxy using Next.js rewrites function.
I have created a reverse proxy like this in next.config.js
async rewrites() {
return [
{
source: "/api/:path*",
destination: `${process.env.NEXT_PUBLIC_BACKEND_URL}/:path*`,
},
];
},
then changed my cookie to be like this
res.cookie("auth_session", token, {
httpOnly: true,
secure: true,
sameSite: "lax",
maxAge: 60 * 60 * 24 * 30,
path: "/",
})
Years later and another solution to limit the width of the ouput of commandlines:
# limit the output to "80" characters:
pstree | sed 's#\(.\{,80\}\).*$#\1#'
Found here: https://supabase.com/docs/reference/cli/supabase-migration-up
supabase migration up [flags]
Flags
--db-url <string> Optional
Applies migrations to the database specified by the connection string (must be percent-encoded).
Put the db-url with (sslmode=require) then we can run migration to remote table which enabled ssl
Simple FILTER
by Team column. Wrap it by CHOOSECOLS
to extract needed columns:
=CHOOSECOLS(FILTER($A$2:$E$4,$E$2:$E$4=B7),1,2,4)
Adapt to your ranges.
This might be helpful
func CalculateRequestSize(req *http.Request) (int, error) {
b, err := httputil.DumpRequestOut(req, true)
return len(b), err
}
func CalculateResponseSize(resp *http.Response) (int, error) {
b, err := httputil.DumpResponse(resp, true)
return len(b), err
}
In this hands-on workshop, you will learn how to upload files to an Oracle Database using only PL/SQL:
https://livelabs.oracle.com/pls/apex/r/dbpm/livelabs/view-workshop?wid=4127
In your case, you are not awaiting the function "fetchData", that is why it s returning only promise and not the data.
Worked me on Dotnet Core 8
on code behind
Description = Description.Replace("\r\n", "<br>");
on razor page
@Html.Raw(item.Description)
For RN Version 0.78.2 and above Add
implementation 'com.facebook.fresco:animated-gif:3.2.0'
in your android/app/build.gradle
In the second schema, the comment field is a required string, where it was a nullable string in the past.
This is a forward compatible change (the old schema can be used to read data written with the newer schema), but not a backward compatible one: the new schema cannot read null values.
To fix, you need to either relax compatibility rules (but that usually causes other problems), or make the field nullable again.
Ok - managed to work it out :-)
Each item on the screen has an attached stylesheet - so we can set the stylesheet of the item e.g.:
self.SongList.setStyleSheet("QListView::item:selected {background-color: #000088;}")
This sets the currently selected item to blue - see https://doc.qt.io/qt-6/stylesheet-examples.html
The trick is to use the "::item:" element...
In the preferences dialog in my app I can allow the user to select what colour they want:
color = QColorDialog.getColor()
# use color.name() to get the #nnnnnn value
Many thanks for the suggestions, however, this allows me to give the user the ability to keep the theme, but then select what they want as a highlight colour.
I've got a few more tweaks and testing to do, but I'll put uploading the code to https://github.com/carlbeech/OpenSongViewer in a few days.
Thanks
Carl.
You can extract text from a PDF file saved as BLOB data in an Oracle DB. Please have a look at:
This issue is occurring due to forgetting some basic checks while installing project dependencies. In my case the main issue was that the TypeScript version did not match the latest version of Angular [19].
Updated following thisngs to run my project.
Update angular/cli from 16 to 19
.
Update angular/devkit to 19
.
Update Typescript 2 to 5.6
.
Before run project check all the version install on your device. if any mistmatch version can caues this issue
hey you guys what's the way to do this in expo snack in mobile as you can't have the access to terminal what's the right version of babel runtime
So, would the reference system have to be specified exactly so that this minimal difference between the raster's extent and cut doesn't occur?
And the other question, I have this example code on a plot with the Terra package of any raster to which I add the extent:
`library(terra)
# Example raster
r <- rast(system.file("ex/elev.tif", package="terra"))
# Original extent
ext_original <- ext(r)
# Expanded extent (larger on all sides)
ext_expanded <- ext(
xmin(ext_original) - 1,
xmax(ext_original) + 1,
ymin(ext_original) - 1,
ymax(ext_original) + 1
)
# Expanded raster (with NA in the new areas)
r_expanded <- extend(r, ext_expanded)
# Plot 1: original raster
plot(r, main = "Original raster with normal extent")
# Plot 2: expanded raster (larger, with NA on the edges)
plot(r_expanded, main = "Raster with expanded extent")
`
This is the raster plot without the background map.
Now to this map I want to add the "background map" from the previous code of the maptiles library but keeping the same format of the raster map with the extension, but it happens that it does not respect the format of the terra plot itself.
`library(terra)
library(maptiles)
# Example raster: global elevation
r <- rast(system.file("ex/elev.tif", package="terra"))
# Original extent and expanded extent
ext_original <- ext(r)
ext_expanded <- ext(
xmin(ext_original) - 1,
xmax(ext_original) + 1,
ymin(ext_original) - 1,
ymax(ext_original) + 1
)
# Download OSM tile with the expanded extent
osm_rast <- get_tiles(ext_expanded, provider = "OpenStreetMap", crop = TRUE, zoom = 8)
# Plot base map
plot(osm_rast, axes=FALSE, legend=FALSE)
# Plot raster on top
plot(r, add=TRUE)
` and this is the map with the "openstreetmap style background map" My question is how can I make the OpenStreetMap style map be added to the background without changing the format of the map plot without a background with the Terra library?, that is, the same map without a background with the extension, but with the background, I don't know if I'm making myself clear.Because as you can see the difference is huge and as you can see in image 2 the latitude limits seem to be "not respected"
If there is still interest in this subject, here is a hands-on tutorial and sample code showing how to achieve it:
Hands-on reading .docx and .pdf file using PL/SQL
Please share your feedback. Thanks.
vim.o.termguicolors = true
Adding this to init.lua file solved my problems.
The "sout" shortcut is an IDE feature, not a Java language feature. It's most commonly associated with IntelliJ IDEA, but similar shortcuts exist in other IDEs like Eclipse or NetBeans. This shortcut doesn't "appear" in a public class because it's not actual Java code - it's just a typing shortcut that your IDE expands into the full System.out.println() statement when you type it. If you're not seeing the shortcut work:
Make sure you're typing it inside a method body, not at the class level Check that code completion/templates are enabled in your IDE Try pressing Tab or Enter after typing "sout" to trigger the expansion
I like @sonjz answer (https://stackoverflow.com/a/24190283/4582204) the best, but it is case-SENSITIVE. To make it case-INsensitive, here are two options (edited repost of: https://stackoverflow.com/a/79626135/4582204).
mode modifier (?i)
:
([regex]'(?i)a.b').Matches('A1B_A2B') | ft
Or use the class function, rather than instance/object function, and pass an option or flag.
$string = 'A1B_A2B'
$regex = 'a.b'
$flags = 'IgnoreCase'
[regex]::matches($string, $regex, $flags) | ft
or simply:
[regex]::match('A1B_A2B', 'a.b', 'IgnoreCase') | ft
Both approaches return a 2-element array (really a MatchCollection
):
Groups Success Name Captures Index Length Value
------ ------- ---- -------- ----- ------ -----
{0} True 0 {0} 0 3 A1B
{0} True 0 {0} 4 3 A2B
I think the most convenient answer (besides the mode modifier (?i)
above) which is case-insensitive and also will match multiple times on a single line (-Match
will not) (https://stackoverflow.com/a/24190283/4582204) is:
$string = 'A1B_A2B'
$regex = 'a.b'
$flags = 'IgnoreCase'
[regex]::matches($string, $regex, $flags) | ft
or simply:
[regex]::match('A1B_A2B', 'a.b', 'IgnoreCase') | ft
Which returns a 2-element array (really a MatchCollection
):
Groups Success Name Captures Index Length Value
------ ------- ---- -------- ----- ------ -----
{0} True 0 {0} 0 3 A1B
{0} True 0 {0} 4 3 A2B
Look at the PHP module BCMath BC Math
It uses strings as numbers with any length and have all common mathematical functions to operate with them.
<?php
$a=bcadd('1234567890123456789012345678901234567890','1000000000000000000000000000000000000000');
echo $a;
Use the Preferences: Open Default Keyboard Shortcuts (JSON)
command in the command palette.
Check for env variables, remove them then also remove the env configuration you have in your babel.config.ts after that stop the app and run npx expo start -c. That worked for me
public String altPairs(String str) {
String s = "";
for(int i = 0; i <= str.length(); i+=2)
{
if(i%4 == 2){
s+= str.substring(i-2,i);
}
if(i%4 == 0){
if(i == str.length()-1){
s+= str.substring(i);
}
}
}
return s;
}
simple way
extension=mbstring
php.ini
I had to install nodejs :D
sudo apt update
sudo apt install nodejs npm -y
everything works fine now
If there is ( @plugin "daisyui"; ) written before, then remove it. It should work.
were you able to resolve this?
so apparently it's called "Components V2" and guess what, it's not available on discord.py, only discord.js
a call for me to migrate into discord.JS
Yep, scratched a bit on that one. There must be no overflow elements, ie the scroller-starter must be within the viewport of the mobile, for it to work, i mean remain static aso that it can trigger animations. Also maybe specify "scroller:window" in scrollTrigger parameters to be sure the mobile stick to it.
Reverting from Twilio.AspNet.Core 8.1.1 to 8.0.2 solved the problem. The issue, as highlighted in https://github.com/twilio-labs/twilio-aspnet/issues/156, is that nullable is now enforced.
I get that AI checkers aren't always correct but come on 3 em dashes and 100% percent AI score on 2 different AI checkers?
If you're going to write thinly veiled rating boosters so people on Google see more results about your company at least hand type them. (And in any case, this is wildly off-topic)
I recently faced this issue when trying to display a PNG from my drawable folder. The problem turned out to be the image size or format, which wasn't compatible with the decoder used by Jetpack Compose/Compose Multiplatform.
By compressing the images, the decoder was able to load them properly, and the crash no longer occurred.
4.5.14 version you can createMinimal and it will suppress the message
CloseableHttpClient httpClient = HttpClients.createMinimal();
For DRF, it is better to use the SessionAuthentication
from rest_framework.authentication import SessionAuthentication
class ApproveOrDeclineUserView(APIView):
authetication_classes = (SessionAuthentication, )
Or use directly for whole endpoints:
# In your DRF settings
REST_FRAMEWORK = {
'DEFAULT_AUTHENTICATION_CLASSES': [
'rest_framework.authentication.SessionAuthentication',
]
}
And don't forget to remove decorator - @method_decorator
I got a reply from Microsoft support that "functionAppScaleLimit" is not available for python flex consumption plan.
Does anyone else have any suggestion on how it can be debugged? Right now I have disabled sampling so all logs are traced, but even then the function just silently stops sometimes without emitting any error. I have try-except blocks everywhere with timeouts implemented on each request...
Import SSL
and export the certificate from the browser
context = ssl.create_default_context(cafile='www.ebay.com.pem')
print("Opening URL... {}".format(context))
This is an alternate option that worked for me.
The original Python Package site: https://pip.pypa.io/en/stable/installation/ specifies multiple options.
One of them is to use the get-pip.py script - https://pip.pypa.io/en/stable/installation/#get-pip-py. Once we download the script, use the following command (python or python3).
python get-pip.py --break-system-packages --trusted-host pypi.org --trusted-host files.pythonhosted.org
Without the argument "--break-system-package", it was giving another error which was addressed in another stackoverflow - How do I solve "error: externally-managed-environment" every time I use pip 3? . I used it any way accepting the risk as its isolated to pip. So, use it with caution.
Without the argument "--trusted-host", there is an SSL cert issue and that is addressed in the stackoverflow - pip3 Could not fetch URL https://pypi.org/simple/pip/: There was a problem confirming the ssl certificate
Thank's to cHao the answer is found.
On the Main Form that you want to branch out from, add this to your buttons event handler (The Word Custom should be replaced with the form you wish to go to.)
Custom^ newForm = gcnew Custom();
this->Hide();
newForm->ShowDialog();
this->Show();
delete newForm;
`
And on the subform you should include this line in another buttons event handler (I am using it in a back button.) This returns you back to the original form.
this->DialogResult = System::Windows::Forms::DialogResult::OK;
I had problems with wgrib2 accepting negative numbers, so I had to make adjustments as follows:
Dataset: GFS 0.25 Degree Hourly
# bounding box for domain
leftlon: -162.
rightlon: -80.
bottomlat: 24.
toplat: 65.# what values to use for wgrib2 slicing?
leftlon positive: -162+360=198
rightlon positive: -80+360 = 280
(rightlon+360)-(leftlon+360)=280-198=82
82*4 =328
toplat-bottomlat = 65-24=41
(toplat-bottomlat)*4 = 41*4=164-lola (leftlon+360):[(rightlon+360)-(leftlon+360)]*4:(degree)
(bottomlat):(toplat-bottomlat)*4:(degree)
-lola (-162+360):[(-80+360)-(-162+360)]*4:0.25 (24):(65-24)*4:0.25
-lola 198:[(280)-(-198)]*4:0.25 (24):(65-24)*4:0.25
-lola 198:82*4:0.25 24:41*4:0.25
-lola 198:328:0.25 24:164:0.25# -lola out X..Z,A lon-lat grid values X=lon0:nlon:dlon Y=lat0:nlat:dlat Z=file A=[bin|text|spread|grib]
wgrib2 gfs_2025051600f000.grib2 -lola 198:328:0.25 24:164:0.25 output.grb grib# without degree 0.25, using degree 1.0 - loss of resolution!
#wgrib2 gfs_2025051600f000.grib2 -lola 198:82:1 24:41:1 output.grb grib
I was able to verify the results in xygrib
note: technically for GZIP this is Z_BEST_COMPRESSION. not hardcoded to 9.
If your AndroidManifest.xml file exists, just go through it; you might have repeated something in it.
I had the exact issue and did online search and found a solution as shared by @mijen67. Its 2025 and it seems VS code came out with what i refer as a straight forward solution with less typing in their own link as shared below
https://code.visualstudio.com/docs/cpp/config-mingw
it will definitely save you several steps as shared by mijen67 in the first solution
I don't know if the problem because your dev server or not, but I recommend to use FlyEnv as your local development server. You can download it from here
all preds must be between 0 and 1 ( 0 < pred < 1 )
for the cost function not to return infinity or nan
From near the top of the page:
Webhooks currently notify you about changes to pages and databases — such as when a new page is created, a title is updated, or someone changes a database schema. The events themselves do not contain the full content that changed. Instead, the webhook acts as a signal that something changed, and it’s up to your integration to follow up with a call to the Notion API to retrieve the latest content.
The body of the webhook request will not contain any specific details about the page created, it will just list relevant information related to the creation event. Who did it, when it happened, etc. If you want more details like title or custom properties of the page created, you should use the ID provided by the webhook payload to look up that information separately.
I was able to find a solution by adjusting the logging level of "Microsoft.Identity". This worked while "MSAL.NetCore" and "Microsoft.Identity.Client" did not seem to be keys that resulted in any real adjustment.
.MinimumLevel.Override("Microsoft.Identity", LogEventLevel.Warning)
Eu também estava enfrentando esse problema. O que funcionou para mim foi não modificar o repositório diretamente pela criação no GitHub, pois qualquer alteração inicial, como adicionar um arquivo LICENSE ou README.md, pode causar o erro "Push Rejected".
Pelo que percebi, o aplicativo SPCK Editor no Android não funciona bem se houver qualquer modificação prévia no repositório remoto.
Solução:
Crie um repositório no GitHub sem nenhum arquivo inicial (sem README, LICENSE ou .gitignore). Depois disso, clone o repositório no SPCK Editor e faça o push normalmente.
Se quiser adicionar um LICENSE ou README.md, baixe esses arquivos de outro repositório e adicione manualmente ao seu projeto no SPCK antes de fazer o push. Assim, evita problemas de sincronização.
The recursion can be defined with @typedef tag.
/** @type {<T>(v: T)=>T} */
const h = (v) => v;
/** @typedef {{bar: number, foo: recursive}} recursive */
const a = h(
/** @returns {recursive} */
() => {
return {
bar: 2,
foo: a()
};
}
);
Yes, it can be using subpath while mounting a volume subdirectory to a container. However,
volume must exist beforehand and
subdirectories must be created before containers attach to it
To create a volume with subfolders, I made a utility function for me to use. I do:
# source the function
source create_volume_with_folders.sh
# create a new volume with however many folders you want
create_volume_with_folders home_data pgadmin pgdata
After that, pay attention to setting external=True
for the volume in docker compose file. See from my create_volume_with_folders.sh Gist.
I started with the pdf-parser-client-side
index.js
file and modified it as below. From K J's answer above, I found that each item has a transform array, and that its 5th element increases with each line. I used that to insert any array that could late be used to split the data into the lines I needed.
'use client';
import { pdfjs } from 'react-pdf';
pdfjs.GlobalWorkerOptions.workerSrc = `//unpkg.com/pdfjs-dist@${pdfjs.version}/build/pdf.worker.min.mjs`;
async function extractTextFromPDF(file, variant) {
try {
// Create a blob URL for the PDF file
const blobUrl = URL.createObjectURL(file);
// Load the PDF file
const loadingTask = pdfjs.getDocument(blobUrl);
const pdf = await loadingTask.promise;
const numPages = pdf.numPages;
let extractedText = '';
// Iterate through each page and extract text
for (let pageNumber = 1; pageNumber <= numPages; pageNumber++) {
const page = await pdf.getPage(pageNumber);
const textContent = await page.getTextContent();
let transform = textContent.items[0].transform[5];
let pageText = [];
// insert '*!*' each time transform changes to separate lines
for (let i = 0; i < textContent.items.length; i++) {
const item = textContent.items[i];
if (item.transform[5] !== transform) {
transform = item.transform[5];
pageText.push('*!*');
pageText.push(item.str);
} else {
pageText.push(item.str);
}
}
pageText = pageText.join(' ');
return pageText;
}
console.error('Error extracting text from PDF');
// Clean up the blob URL
URL.revokeObjectURL(blobUrl);
} catch (error) {
console.error('Error extracting text from PDF:', error);
}
}
export default extractTextFromPDF;
Examples:
textContent.items[2] = {str: 'Friday', dir: 'ltr', width: 21.6, height: 6, transform: Array(6), …}
textContent.items[2].transform = [6, 0, 0, 6, 370.8004000000002, 750.226]
pageText = 'Produced Friday 09/13/24 14:22 Page No. 1 YYZ *!*...'
I am having this exact problem with my Flask React app. It works perfectly fine when I test it locally and when I run my Docker container locally. But for some reason, when I run the Docker container in my VPS, I get nothing. It has to be something with my NGINX config and my reCAPTCHA settings. It's gotta be! lol
df.filter(pl.col.a == pl.lit(['1']))