From comment by Jonathan Leffler:
Use unsigned char for the front and rear indexes — that will avoid the warning safely.
Nailed it. Built all the way through that time.
// FIFO queue structure for tracking the keys
typedef struct {
char items[6];
unsigned char front;
unsigned char rear;
} Queue;
I'm not entirely sure about the solution, but I found a package that claims to support NFC with Expo.
You can check out the Expo NFC Package. Give it a try, and if it works, that's great! If not, feel free to reach out, and we can troubleshoot it together.
I was also facing this problem with many of my websites where i use videos. These days many videos are as reels i.e. vertical or portrait.
I tried many option like mentioned above. They work but for the sake of simplicity i created a plugin to make things simple. For people who may not be very comfortable with css or inline code.
This plugin is free and on wordpress. https://wordpress.org/plugins/yt-portrait-video-embed/
Simply use shortcode and video ID. And its done. I have tried to make it responsive as well. Hope its helpful to you.
######ABOVE######
Nailed it!
const date = new Date() // current date
const minutes = date.setMinutes(date.getMinutes() + yourMinuteshere)
const hours = date.setHours(date.getHours() + yourHourshere)
You can add viewbinding to avoid findViewById , check official ViewBinding document.
Having the same error, do you have this solved already?
It is a compatibility issue between Avro and Spark. Remove from the spark.jars.packages, org.apache.spark:spark-avro_2.12:3.5.3 since in the Spark latest vesrion already included. Secondly use latest Hadoop-Azure version "org.apache.hadoop:hadoop-azure:3.3.4". Refresh your spark session, it will resolve the issue.
I've fixed the issue. The issue was actually with the material CSS file imported in angular.json .
If we want to use old styles we need to use this line - ./node_modules/@angular/material/legacy-prebuilt-themes/legacy-indigo-pink.css
NOTE: This can be used till v16 only
If we are talking about the same package speech_to_text, maybe there's no way. It wraps all native systems voice recognition, so there's not a direct way to customize the dictionary. For example:
https://developer.android.com/reference/kotlin/android/speech/SpeechRecognizer https://developer.apple.com/documentation/speech
Analyzing both Android and iOS APIs, they are very limited to download languages and in most cases, infering in some results like Android AlternativeSpan, but I don't think that's the case.
I don't know if you are planning to use only local or cloud Speech to Text recognition, but I would recommend Google Speech-to-Text API.
It's available on googleapis package.
I've also searched for some edge models and could only find https://huggingface.co/openai/whisper-small as tiny version, but not sure how "tiny" it is for a mobile phone.
Why dont you just execute hot reload, hot restart or even build your projects in windows if your windows have better specifications?
Also, what do you mean by overwrite each other, maybe you can explain more?
Happened to me as well and then I had it resolved by emailing their support team, Here's their email. Heroku [email protected]
There are many other answers on Stackoverflow for this error message. In short, modules that use C or C++ code are compiled against the version of perl you are using and have markers in them to check they are being used with that version. If you change the perl version (or compiler details in some cases), those won't work and you get that message.
Likewise, if you update the shared library that one of those compiled modules wants to use, you'll have a similar problem. For example, I have to keep around the openssl libraries I compiled against for a module that used those to keep working. If I update openssl (at the same location), I need to recompile the Perl modules that work against those.
CPAN.pm can recompile everything it knows you have installed:
cpan -r
Mostly this sort of thing happens when you are trying to share a module library outside of the default paths (say, like using local::lib). Reinstall those modules too.
And, don't share that directory between different versions of perl, such as the system perl and one you installed through a different method.
You can use the limit parameter in your API call but you can still fetch only 100 records at a time and you'll have to paginate post that.
Curious to know what your use case is! Are you planning to regularly sync data between Chargebee & Power BI?
Another thing, if you're having issues with dependencies, and if you're using Visual Studio Code.
Try to check the OUTPUT tab right next to PROBLEMS (or you can using shortcut commands: Ctrl+Shift+U) There are information about plugins' version you've been using, whether they're compatible with each other or not. It will update the information after by having save changes in the dependencies of pubspec.yaml file.
It is useful for me to understand the compatibility between my project and the plugins dependencies itself.
Got the same problem. If you run
!pip install -U git+https://github.com/google-gemini/generative-ai-python@imagen
import google.generativeai as genai
import pprint
for model in genai.list_models(page_size=100): # page_size can be any number
pprint.pprint(model.name)
Available Models Instead of seeing the Imagen 3 model ("imagen-3.0-generate-001") or any image generation models, here's what it returns, or at least what I get:
Problem: According to Google's blog post, there is a waitlist for accessing Imagen 3. To join, you need to fill out the "Labs.google Trusted Tester Waitlist," which requires providing credentials—a process that feels like getting military clearance.
Has anyone successfully accessed Imagen 3 through the API? Any tips or alternative steps to get this model up and running?
enter code hereI encountered a similar issue and resolved it by adding multiple links to the configuration file located in public/env/config.js (and in config.testing.js).
To add multiple links to the same environment, you can specify them in the VITE_APP_UI_URL variable, separating each URL with a comma. This approach allows you to easily configure multiple links within the same environment.
//config.Testing.js
process.env.PORT = 3000;
process.env.HOST = 'localhost';
process.env.CLIENT_ID = 'your_client_id';
process.env.REDIRECT_URI = 'your_callback_endpoint';
var VITE_APP_UI_URL = "your_test_url_1 , your_test_url_2";
var VITE_SITE_KEY = "your_site_key";
var VITE_APP_API_BASEURL = "your_baseurl";
I had the same issue, but it was resolved in the latest version, 17.11.15, of Visual Studio. I recommend running a repair on the IDE through the installer to fix it.
The problem is that you are creating one chart for each entry of your data set.
Create a single ColumnSeries for your entire dataset.
Another thing, if you're having issues with dependencies, and if you're using Visual Studio Code.
Try to check the OUTPUT tab right next to PROBLEMS (or you can using shortcut commands: Ctrl+Shift+U) There are information about plugins' version you've been using, whether they're compatible with each other or not. It will update the information after you execute,
flutter pub get
It is useful for me to understand the compatibility between my project and the plugins dependencies itself.
Well, the code you provide is working fine, It wont create
RenderViewport#252ca NEEDS-LAYOUT NEEDS-PAINT NEEDS-COMPOSITING-BITS-UPDATE:
unless you put the ListView.builder inside the Column or Row widget
For Xcode 15.0+, you have to remove the Main interface configuration from here, as attached:
Steps:
Note: you also need to make the changes in info.plist file related to Main interface. That is same process as the previous Xcode versions.
Cant be done. Solution was add defaults of empty string to the columns in the table.
For anyone who see this question later, I have developed a Kotlin/JVM library to facilitate working on Hijrah dates and times, which is built on top on java.time.chrono.
It adds a set of new classes, and extention functions to current java classes, include functions like plusWeeks, plusMonths, publicly available. Give it a try!
You can find the solution here
I add the key file content as env var to cloud function and followed instructions in the answer below
Google OAuth using domain wide delegation and service account
thank you so much for sharing this piece of content, I was also looking for the solution of BaoBao Technology website
When placing a swiper-container inside a mat-grid-tile within an Angular Material mat-grid-list, the swiper component doesn’t display correctly by default. The slides either don't take up the full space, or they have layout issues.
To fix this, add width: 100% to the swiper-container. This will make it take up the full width of the mat-grid-tile, allowing the swiper to render and distribute the slides correctly.
swiper-container { width: 100%; }
I’m using FVM to switch from SDK 2.10.5 to 3.24.2, but I get this error:
Future isn't a type. Try correcting the name to match an existing type.
Initially, I tried importing dependencies and updating the SDK environment in pubspec.yaml in VSCode, but nothing worked. Then, I opened the same project in Android Studio Koala (the newer version), and it ran successfully without any code changes.
command I used in Android Studio(already select sdk using fvm):
fvm flutter pub get
and
fvm flutter run
So, I can assume this issue might be caused by the IDE or Flutter/Dart settings in the IDE(both IDEs have the same SDK path and the same versions of Flutter and Dart installed).
Please add a trailing slash at the end of proxy_pass URI as http://rails_app:4000/
Adding the trailing slash tells NGINX to strip the /api prefix from the request URL before passing it to the backend service. This way, if a client requests /api/posts, it will be proxied to http://rails_apps:4000/api
When you are in your report design window, select the tablix.
At the bottom, you will see Row Group to your left and Column Group to your right.
At the very edge of Column Groups, there is a tiny downward pointing dropdown, click on that and select "Advanced Mode"
That will open you your Static Groups under your Row Groups.
Select the first one, and then go to properties. Set RepeatonNewPage = False
I’m encountering the same issue as others where my Angular application takes up to 5 minutes to load when generating a PDF using either Puppeteer or Playwright. The problem seems to be related to the main.js file of my frontend application, as it is the only file that takes a long time to load.
I initially thought it could be related to the Docker image I was using. When I used my local Docker image, I was able to generate the PDF successfully. However, when I deployed the image to Rancher, the PDF generation failed, and it took up to 5 minutes to load the main.js file.
I’ve tried various solutions, including switching to the Node Slim image and alternating between Puppeteer and Playwright, but neither solution resolved the issue.
Has anyone encountered this problem or know what could be causing the long load time for main.js when generating the PDF, specifically in the Rancher environment?
Any insights or suggestions would be greatly appreciated!
Issue with sending emails through Gmail’s SMTP server. Check Server Firewall and Outbound Connections. Check PHPMailer's Debugging Output. Update PHP Version and Extensions. Testing with Another SMTP Server. By enabling PHPMailer's debug output, checking server restrictions, and ensuring your PHP configuration is correct, you should be able to get more details on what's going wrong.
you can run swagger to look up the segment id and then return to you all mcsvids. this is useless however when speaking with other non-adobe-marketing tools. but you could return other things like userids or hash ids.
Just found a simple answer using the "+" operator
myText = "Number is "
myText = myText+ 10
if you initiated OneDrive, most of your files would most likely have admin in it's path. Also when you open jupyter notebook, you would experience errors such as this : Your file couldn’t be accessed It may have been moved, edited, or deleted. ERR_FILE_NOT_FOUND
To solve this, first, open file explorer and paste this directory C:/Users/Admin/AppData/Roaming
look for jupyter folder, if it exist, delete it. Next, create another jupyter folder and open it the create a runtime folder
Close Anaconda prompt or any terminal .
reopen your anaconda prompt or cmd using run as administrator and type jupyter notebook
You should be able to open jupyter notebook in the web and access all your files including OneDrive
I don't know if this is a valid aproach or not, but you can try change tintColor of PHPickerViewController:
picker.view.tintColor = .black
Guys i written a code for merging file but i want to re modify it but don't have proper idea can anyone help me?
Sub Master_File()
Dim n As Integer
Dim wb As Integer
Dim master As String
Dim userfile As String
'Dim l_Row As Long
'Dim l_Dist As Long
Application.DisplayAlerts = False
Application.ScreenUpdating = False
master = ThisWorkbook.Name
With Application.FileDialog(msoFileDialogOpen)
.AllowMultiSelect = True
.Title = "Locate Your Files"
.Show
n = .SelectedItems.Count
For wb = 1 To n
Path = .SelectedItems(wb)
Workbooks.Open (Path)
userfile = ActiveWorkbook.Name
For Each Sheet In ActiveWorkbook.Worksheets
If Sheet.Name = "Tracker" Then
Sheet.Select
Sheets("Tracker").Select
Range("a1").Select
Range(Selection, Selection.End(xlToRight)).Select
Range(Selection, Selection.End(xlDown)).Select
Selection.Copy
'l_Row = Sheets("Tracker").Range("A1048576").End(xlUp).Row
'This will find the last row of the tracker sheet
'Range("A2:K" & l_Row).Copy
'This code will copy all dat from tracker sheet
Windows(master).Activate
'This code will activate the master file where we will paste our data
Sheets("Sheet1").Select
Range("a1").Select
'Range("a1048576").Select
Selection.End(xlDown).Select
Selection.End(xlDown).Select
Selection.End(xlUp).Select
ActiveCell.Offset(1, 0).Select
'l_Dist = Sheets("Sheet1").Range("A1048576").End(xlUp).Row + 1
'This code will find the next blank row in the master file
'Sheets("Sheet1").Range("A" & l_Dist).Select
'This code will find the last non blank cell in the master file
Selection.PasteSpecial Paste:=xlPasteValues, Operation:=xlNone, SkipBlanks _
:=False, Transpose:=False
Range("b1048576").Select
Selection.End(xlUp).Select
Selection.End(xlToLeft).Select
ActiveCell.Offset(1, 0).Select
Range(Selection, Selection.End(xlDown)).Select
Range(Selection, Selection.End(xlToRight)).Select
Selection.Delete
Range("a2").Select
Selection.AutoFilter
ActiveSheet.Range("$A$2:$L$41").AutoFilter Field:=1, Criteria1:= _
"Employee Number"
ActiveCell.Offset(20, 0).Range("A1").Select
Range(Selection, Selection.End(xlToRight)).Select
Range(Selection, Selection.End(xlDown)).Select
Selection.EntireRow.Delete
ActiveCell.Offset(-20, 0).Range("A1").Select
Selection.AutoFilter
Range("a2").Select
Range(Selection, Selection.End(xlToRight)).Select
Range(Selection, Selection.End(xlDown)).Select
Selection.Font.Bold = True
Selection.Columns.AutoFit
With Selection
.HorizontalAlignment = xlCenter
.VerticalAlignment = xlBottom
.WrapText = False
.Orientation = 0
.AddIndent = False
.IndentLevel = 0
.ShrinkToFit = False
.ReadingOrder = xlContext
.MergeCells = False
End With
ActiveCell.Select
Windows(userfile).Close savechanges:=False
End If
Next Sheet
Next wb
End With
End Sub
In this code i want to edit at here like i want to connect the sheet name from cell A1 like what i write in cell A1 that will be consider as sheet name from other workbooks for find the sheet and then it will find that name from other workbook sheets if i write in cell A1 = Company then it will find Company sheet name from other workbook sheets
**
If Sheet.Name = "Tracker" Then
Sheet.Select
Sheets("Tracker").Select
Range("a1").Select
**
Can anyone help me Please?
Extract the header from the list and add it as a new row.
header_row <- colnames(scraped_data[[1]])
scraped_data[[1]] <- rbind(header_row, scraped_data[[1]])
Note that the column name remains the same
As @regdos correctly pointed out, do a copy rather than an update -- this combined with memo actually ended up saving the render of the second ContentComponent.
Flamegraph showed Did not render on the client during this profiling session. which is what I wanted!
I was able to resolve this error after updating beam version to 2.59 and installing apache-beam[gcp].
After disconnecting from VPN, I was able to install this package from PyPi repo.
pip install apache-beam[gcp] --index-url=https://pypi.org/simple
It turns out that ghcr.io uses basic authentication in the form of <username>:<PAT>
So it would look something like this.
curl -H "Authorization: Basic $BASE64_CREDENTIALS" "https://ghcr.io/token?service=ghcr.io&scope=repository:<user>/<image>:pull"
{"token":"xyz"}
I think onload function with state can work also skeleton component need some css properties or their inbuilt settings.
Or, maybe you could use some CSV reader library to looping through the CSV contents like this?
const fs = require('fs');
const csv = require('csv-parser');
fs.createReadStream('test.csv')
.pipe(csv())
.on('data', (row) => {
console.log(row['Modem ID'], row['Total Chargeable'].replace(',', '.'));
})
.on('end', () => {
console.log('CSV file successfully processed');
});
I think this approach would be better, rather that you read the CSV as a string
Velero has started to deprecate Restic as documented here.
Old answer, but I've made a VM detection library in C++ that is meant to do exactly this.
https://github.com/kernelwernel/VMAware
All the techniques mentioned in this post are included in the library.
First, you can replace "," to another character. For example, you can change "BF002062010","1,300.00" to "BF002062010"-"1,300.00".
Then, using line = line.replace(',', ''); removes thousand comma.
Finally, you can replace "-" to replace ",".
Add back the comma.
line.replaceAll(',', '').replace('""', '","')
location: {
displayName: <your Room name>,
locationEmailAddress: <your Room emailAddress>,
locationType: 'conferenceRoom'
}
this's work for me
The outliers are so far away that it forces the box plot to be squished as the range of data is so large, you could use outlier detection techniques to help with this.
Using a log(y) will help by reducing the amplitude of the outliers and the spread of values and make the box plot more easily visible.
You can also click the zoom button above the plot in Rstudio if you are using that as your IDE, this will open the plot in a new window that can be fullscreen.
Hope this answers your question.
Does this do what you're looking for?
df$Percentage <-
rowSums(df[as.character(yrs)]) /
rowSums(
outer(df$CofQYr, yrs, FUN = function(x, y) y <= (yrs[length(yrs)]-as.integer(x))) *
as.matrix(df[as.character(yrs)])[match(df$TRADE, ifelse(df$CofQYr=="Total", df$TRADE, NA)),]
)
Turns out I just needed to do a little more reading. os._exit(0) is what works.
import time
import keyboard as kb
import sys
import os
def on_hotkey():
print('hotkey')
os._exit(0)
kb.add_hotkey('ctrl + q', on_hotkey )
for i in range(3):
print('Hello')
time.sleep(5)
this is consistent with tidyselect and works within groups:
filter(c_across(all_of(var_a))==4)
These are the CSP directives that Stripe.js requires https://docs.stripe.com/security/guide?csp=csp-js#content-security-policy
Can you add them to your integration and try again?
You could refer to following url to filter by date
https://domain.sharepoint.com/sites/sitename/_api/web/lists/getByTitle('listname')/items?$filter=Date ge datetime%272017-19-05%27
How to make a chat bot though and integrate it like so it doesn’t have a bot tag and can talk to people in dms?
this might be consistent with tidyselect and works within groups:
filter(!is.na(c_across(all_of(var_a))))
You can see all the supported browsers by:
npx playwright install --help
and then download a specific browser by:
npx playwright install chrome
To interact with shadow dom with puppeteer. Keep everything same like normal dom except selectors.To Handle element from shadow dom add ">>>" before selector.
For example
await page.click("selector")//for regular dom
await page.click(">>>selector")//for shadow dom
For inline, us the brackets sign like the example below.
Hello, world!I found out why. Our service was hosted on the isolated network only accessible by office wifi.
The tester was using the 5G network when he tried to make request which obviously failed. The reason why other older test devices were working was that they can only use wifi internet as they were not wired to any carrier services.
It wasn't neither of front-end or back-end problem and one of the wierdest error I got haha.
I found out that this problem was reported to Microsoft a few months ago. They said that it had been fixed in the latest version of Visual Studio. I already had the latest update, but it still did not fix it. So I deleted VS, cleaned my registry, rebooted my computer, downloaded/reinstalled VS (same version) and now the problem is gone. So having the latest update might not fix it. Delete, clean registry and reinstall VS was my solution
If you add limit=250000 to your query do you get all of the data? I presume when you say the results are reduced you mean the date range is reduced and not the number of rows returned.
Data from the API is paginated with a default of 10,000 rows and a max of 250,000 rows. For more than 250,000 rows you'll have to also set an offset parameter and run the query again.
You can find details here:
RunReportRequest(
property=f"properties/{GOOGLE_ANALYTICS_ID}",
return_property_quota=True,
dimensions=[
Dimension(name="date"),
Dimension(name="customEvent:itemId”),
# Dimension(name="customEvent:companyId), # Uncommenting causes problem
],
metrics=[
Metric(name="eventCount"),
],
date_ranges=[
DateRange(start_date='2024-03-25’, end_date='2024-09-25’),
],
limit=250000
)
Setting the HorizontalScrollBarVisibility to Auto or Visible should help.
Consider using JsPrettier, which is a front end for Prettier. Prettier is a modern code formatter widely used and actively developed. Besides HTML, it supports a whole lot of other languages. JavaScript runtime is required for it to run (Node.js or Deno).
That's great. How would it need to be to work on iterm instead of Terminal?
I would also verify the connection assignment on the terminal, setting it to "Continuous Output" will give you a constant stream of weight data.
FIXED: The problem was due to vercel. Hosted the server on aws-eleasticbeanstalk and everything worked smoothly.
A potential solution for this error in your Spring application running on Docker is to add the following variable on Dockerfile.
SPRING_MAIN_ALLOW_BEAN_DEFINITION_OVERRIDING=true
Thanks you @dbugger for pointing me in the right direction. @Edmund's answer to this question was the solution to my issue:
https://stackoverflow.com/a/78882862/5161457
Remove the default site if it exists:
sudo rm /etc/nginx/sites-enabled/default
Test your Nginx configuration:
sudo nginx -t
If the test passes, reload Nginx:
sudo systemctl reload nginx
Then retry in 10 mins.
This is probably due to Safari's handling of base64-encoded images and of xlink:href.
Replace xlink:href with href in the <use> elements:
<use href="#image0_1330_2263" transform="translate(-0.43633) scale(0.000749064)" />
The reason for the unauthorized error is because nginx uses a user called www-data.
It is also possible to add permissions without changing the user name, but There is also an easy way to change the user name used by nginx.
How to change nginx username:
$ sudo nano /etc/nginx/nginx.conf
# change user www-data to ubuntu
$ sudo systemctl restart nginx
Since adding only permissions is complicated, it is replaced with a link.
In short, I think that the answer depends on the environment...
I'm recommending this package: https://github.com/single-spa/import-map-overrides
It works for browsers and NodeJS too...
You register your aliases in the main.js (or your entry point), then any other module can use that alias to import the aliased file.
Why the Html.Raw calls on all the strings?
I’m looking at it on my phone but I can’t see why that wouldn’t work as just
class=“something @(hasMM ? “Hey” : “”)”
?
Thanks to Joe for his comment: use numKeys = bson_num_keys(bsonArray) to get the number of keys (entries), and use bson_append_document(bsonArray, "numKeys", -1, bsonObj)
function cikti = sigmoidDerivative(x)
cikti = sigmoid(x) .* (1 - sigmoid(x));
this code part have to be changed to
function cikti = sigmoidDerivative(x)
cikti = x .* (1 - x);
because in the jacobian function, sigmodiDerivative() funtion is called with parameter that already calculated with sigmoid() function
[~ , jA1] = forward_pass(jinput_vals(i),W1,b1,W2,b2);
derivW1 = -1 * W2'.* sigmoidDerivative(jA1) *jinput_vals(i);
jA1 is already return from sigmoid() function and so repeated unnecessary sudden increase and decrease of lamba problem solved.
I am using Docker with a nodejs alpine image. I am trying to install a command line interface for TestRail and this package is a dependency for my tech stack. Running apk add py3-ruamel.yaml solved my issue.
Page for reference
I tried all the css suggestions without any luck. After some experimentation I found that if you replace the html element directly under the element being scaled with itself everything is crisp. Here is a react hook example:
const useChromeScaleHack = (ref: MutableRefObject<HTMLElement>, scale: number) => {
if (!window.chrome) return
useEffect(() => {
if (scale > 1.09 && ref.current) {
const el = ref.current
el.replaceWith(el)
}
}, [scale])
}
const Comp = ({ scale, children }) => {
const ref = useRef(null)
useChromeScaleHack(ref, scale)
return (
<div
style={{
willChange: "transform",
transform: `scale(${scale})`,
}}
>
<div ref={ref}>{children}</div>
</div>
)
}
Ok you can log out of the site and restart it to work or another solution There may be a temporary problem with the site.
I just came across a similar error in my application. I believe it was caused by the encryption key Next uses for server actions cycling for each build.
From the docs: "To prevent sensitive data from being exposed to the client, Next.js automatically encrypts the closed-over variables. A new private key is generated for each action every time a Next.js application is built. This means actions can only be invoked for a specific build."
The docs then go on to say that you can override the process.env.NEXT_SERVER_ACTIONS_ENCRYPTION_KEY variable to keep a consistent server action encryption key across builds.
Read more here https://nextjs.org/docs/app/building-your-application/data-fetching/server-actions-and-mutations#security
Follow this formatting example. The URL should look as below.
{{baseUrl}}/reservations?filters=[{"field":"checkIn","operator":"$between","from":"2023-03-02T00:00:00%2B01:00","to":"2023-03-02T23:59:59%2B01:00"}]
Enjoy
It appears to me that vercel writes the requests to the logs even if the requests return 403 forbidden. That confused me a bit, but they are being denied even though the requests are logged.
At this point I have set up this configuration in the vercel firewall.
Rule 1 RequestPath --> MatchesExpression --> .php$|.php7$
Rule 2 RequestPath --> MatchesExpression -->(wp-content|wp-admin|wp-login|cgi-bin|wp-includes|wp-trackback|wp-feed|.well-known)
I did it this way because the logs show the bots looking for urls like site/wp-content/ or site/wp-includes/ without a php extension just a trailing slash
Today I had 1,000 requests denied with this setup so it seems to be working pretty well.
JA4 Digest requests are next on the menu.
The issue you're facing is due to your trigger call, spinResult.trigger(). The trigger function only takes one argument. If you want to supply multiple parameters to the callback, you can send them in an array.
For example: Here's what the trigger line should look like to rectify this issue:
nc.appEvents.spinResult.trigger( [5, "7777"] ); // pass in the args as an array
As pointed out by @jakevdp, there is an issue with the efficiency in running the function accuracy due to operations that are not JAX native.
Although, I tried to ensure only JAX arrays and fully vectorized operation for compatibility as below;
def accuracy(labels, predictions):
labels = jnp.array(labels)
predictions = jnp.array(predictions)
correct_predictions = jnp.sign(labels) == jnp.sign(predictions)
acc = jnp.mean(correct_predictions)
return acc
but still the performance in evaluating accuracy is very bad, infact it's infeasible.
As I am dealing with Quantum Neural Networks, in specific Parameterized Quantum Ciruits as my model, the number of learnable parameters are very few in number. Thus, a more feasible and efficient workaround would be to log the train and test loss (surprisingly the evaluation of test loss is not very slow), atleast this helps in avoiding overfitting.
Furthermore, for getting the accuracies, save the parameters at each step using jax.debug.callback, and fetch them to calculate the train and test accuracies at each step after the training process.
It can be implemented as below.
import os
import pickle
import time
from datetime import datetime
def save_params(params, step):
# Get current timestamp
current_timestamp = datetime.now().strftime("%Y%m%d%H%M%S")
# Directory where params are saved
dir_path = f"stackoverflow_logs/{current_timestamp}/params"
os.makedirs(dir_path, exist_ok=True) # Create directories if they don't exist
# File path for saving params at the current step
file_path = os.path.join(dir_path, f"{step}.pkl")
# Save params using pickle
with open(file_path, "wb") as f:
pickle.dump(params, f)
print(f"Parameters saved at step {step} to {file_path}")
@jax.jit
def update_step_jit(i, args):
params, opt_state, data, targets, X_test, y_test, X_train, y_train, batch_no, print_training = args
_data = data[batch_no % num_batch]
_targets = targets[batch_no % num_batch]
train_loss, grads = jax.value_and_grad(cost)(params, _data, _targets)
updates, opt_state = opt.update(grads, opt_state)
test_loss, _ = jax.value_and_grad(cost)(params, jnp.array(X_test), jnp.array(y_test))
# Save parameters every step
jax.debug.callback(lambda params, step: save_params(params, step), params, i)
params = optax.apply_updates(params, updates)
def print_fn():
jax.debug.print("Step: {i}, Train Loss: {train_loss}, Test Loss: {test_loss}", i=i, train_loss=train_loss, test_loss=test_loss)
jax.lax.cond((jnp.mod(i, 1) == 0) & print_training, print_fn, lambda: None)
return (params, opt_state, data, targets, X_test, y_test, X_train, y_train, batch_no + 1, print_training)
@jax.jit
def optimization_jit(params, data, targets, X_test, y_test, X_train, y_train, print_training = True):
opt_state = opt.init(params)
args = (params, opt_state, data, targets, X_test, y_test, X_train, y_train, 0, print_training)
(params, _, _, _, _, _, _, _, _, _) = jax.lax.fori_loop(0, 10, update_step_jit, args)
return params
params = optimization_jit(params, X_batched, y_batched, X_test, y_test, X_train, y_train)
var_train_acc = acc(params, X_train, y_train)
var_test_acc = acc(params, X_test, y_test)
print("Training accuracy: ", var_train_acc)
print("Testing accuracy: ", var_test_acc)
Finally, one can also use any frameworks like wandb, mlflow or aim for tracking in the callback function.
Every time I do it, as in unzip the jar and then zip it without any changes, it just says "Error: Invalid or corrupted jarfile" so I don't know if I should give up or not.
there's an official endpoint for this
Use http://api.steampowered.com/ISteamUser/ResolveVanityURL/v0001/?key=YOUR_API_KEY&vanityurl=VANITY_URL
You can get a Steam web api key from https://steamcommunity.com/dev/apikey
Ok, it seems that I found my solution. And it's a classic rookie mistake. The cube I create in Blender is smaller than the one in my own script :D. After adjusting it... it works.
I was able to fix this by removing permissions on the aws service that was being invoked within my recursive function.
I went to IAM and found the role that was being used by the Lambda function. I selected the Permission Policy that was tied to the aforementioned service and clicked "Remove".
I noticed the loop by viewing the Cloud Watch logs of the Lambda Function. After completing the above, the logs stopped with an error about how it no longer had permissions.
This was so helpful. I was struggling with this problem to. And adding this line to style.scss fixed the problem. Thanks.
cwicncslaicnalwcknakociiwiii wcoapwpo ahwc uiu ucoainkpox[nx wccnaocisabc su iawbcsaiwocisna acnoix aiiksncoi[ lokjw uwujaklo iwjidwioi wjxnuwaiwidnxoaiw wuxnanusaiik sisnin owiaiwdn siawdnoaiwdsia iaodwaodsa disaoiwdk siaica woiwahdohsa coaidhsoacno nshliaw swsspaown siwal
I thought it was easy to use,but rejected my email and password together, so understand the situation and let sign in
I'm having similar issue and it happens only with Safari, both Firefox and Chrome do not exhibit this behavior. This means, there must be something related to how Safari accesses the YouTube. I think this behavior can be eliminated by unchecking Safari -> Settings -> Privacy -> Hide IP address from trackers.
This is the normal behavior of Protobuf. Google protobuf library does have a way to overcome this by using JsonPrintOptions and set always_print_primitive_fields to true. I did not find something similar in QT code.
@DazWilkin had it right - I needed to have uint64 instead of int64; now everybody's happy.
Have you found the answer? i have been having the same issue myself of you could help me that would be great thank you.
Create a component with client name on it, TextEditor.client.jsx It will render on client only.
The following css would add a dot to the custom counter:
counter-increment: mycounter;
content: counter(mycounter) ".";
The primary issue is that you're using a shared state object which is set to true and never set to false based on the code you've provided. That means when you first present the sheet, it acts in a way that you're expecting, however the second time you present the sheet, it's only presented briefly then dismissed because you've effectively toggled the boolean to false on that second button. It looks like this, by execution order; Initialized false -> Button Tap -> Set true -> Sheet Presented -> Sheet Dismissed -> Second Button Tap -> Set false -> Weird behavior.
I also noticed you're using $obsPresentation.triggerSheet instead of obsPresentation.$triggerSheet which could also cause the problem, but I recommend below, a better solution regardless. I don't have my IDE right now, so I can't confirm that gut instinct.
So, if you've read this far, then surely, you're wondering, why is the sheet even presenting at all, given that it's set to false. That's because sheets have a unique behavior where they ignore certain thread actions while they're mid-animation. Your implementation is triggering the sheet to appear, because the @State is being updated, and in the context of the ViewBuilder it doesn't care if its true or false, simply that it's updated, and it should re-draw. However, while it's in the middle of that re-draw, it's published value flips to false, thus causing the sheet to disappear. Similar weird behavior can happen by dismissing a parent of the sheet, from the parent, while a sheet is presented as proof of this. So how do you fix it?
The best way to fix a sheet, in particular one that you're using for different things, is to use an Enum in place of a boolean value. For example:
enum PresentedSheet {
case sheetOne, sheetTwo, sheetThree
}
If you were to use an value that is optional, then you can have the sheet be presented, whenever the value is NOT nil. For example:
@State presentedSheet: PresentedSheet? = nil
var body: some View {
Text("Example")
.sheet(item: $presentedSheet) {
switch presentedSheet { //Show the view you want }
}
}
In your case, you'd add the presentedSheet in place of your boolean, on your observable object. If you want additional controls, there is a callback function that can be called onDismiss if you need to know or update something when that sheet is dismissed.
As a final note, when working with sheets, if you're sharing a sheet among different views, be sure to have that sheet presented from the parent most view, or if a particular child view is destroyed it will also automatically dismiss your currently presented sheet. Also, don't overuse this as it can quickly clutter your code or complicate things.