Response
stringlengths
15
2k
Instruction
stringlengths
37
2k
Prompt
stringlengths
14
160
It is easier to start with the recommendation formgitignore.io.Seegitignore.io/api/laravelLaravel/vendor node_modules/ npm-debug.log# Laravel 4 specific bootstrap/compiled.php app/storage/ # Laravel 5 & Lumen specific public/storage public/hot storage/*.key .env.*.php .env.php .env Homestead.yaml Homestead.json # Roc...
I start usingGithub2 days ago, and it's in a private repository.So I with my friend (I appointed him as Collaborator) working a project together, but what is the best setting for.gitignorefile when you're working with a team?Because what I'm thinking is, ifGithubignore file such as.env(the app key) &/vendor, the progra...
Does this .gitignore laravel makes sense when you're working with a team?
Containers isolate applications from each other on the same machine, but you're right, they all use the underlying OS. If you need different OS to run different applications on the same machine, you need to use virtual machines instead. Containers are good because you get everything you need to run an application in ...
As a newbie,I have read the official Docker documentation, and have followed many explanations here, tutorials, videos on this, but have not yet got a clear answer to my question. If a docker container must use the underlying host OS kernel, then how can they claim "build, ship and run anywhere"? I mean, linux-based co...
Are Docker containers tied to the underlying host OS?
Can you share pod's logs?kubectl logs <pod_name>Postgres is using init script with defined variable names:POSTGRES_USER POSTGRES_PASSWORD POSTGRES_DBTry this one outapiVersion: v1 kind: ReplicationController metadata: name: postgres spec: replicas: 1 template: metadata: labels: app: postgres ...
I'm trying to change the settings of my postgres database inside my local minikube cluster. I mistakenly deployed a database without specifying the postgres user, password and database.The problem: When I add the new env variables and usekubectl apply -f postgres-deployment.yml, postgres does not create the user, passw...
Reconfigure postgres with kubectl apply command
Sounds like you're making a backup using the pg_dump utility. That saves the information needed to recreate the database from scratch. You don't need to dump the information in the indexes for that to work. You have the schema, and the schema includes the index definitions. If you load this backup, the indexes will...
When I make a backup in postgres 8 it only backs up the schemas and data, but not the indexes. How can i do this?
How can I backup everything in Postgres 8, including indexes?
I ended up with this solution: you simply start several php-cgi processes and bind them to different ports, and you need to update nginx config:http { upstream php_farm { server 127.0.0.1:9000 weight=1; server 127.0.0.1:9001 weight=1; server 127.0.0.1:9002 weight=1; server 127.0.0.1...
I'm currently usingnginxandPHP FastCGIbut that arrangement suffers from the limitation that it can only serve one HTTP request at a time. (Seehere.) I start PHP from the Windows command prompt by doing;c:\Program Files\PHP>php-cgi -b 127.0.0.1:9000However there is another way to run PHP know as "Fast CGI Process Manage...
Can Windows PHP-FPM serve multiple simultaneous requests?
3 Where does this declaration occur? I think it should fit in the memory of a Linux machine, but probably not on the stack, unless you take special actions (e.g. ulimit -s). In general, it's not a good idea to use large local C style arrays—in fact, except in special case...
How can I increase the memory limit for a C Program. I am using code blocks and trying the following code - int arr[10000000] It is giving me run-time error. I am using Linux(Fedora). Any help...?
code blocks memory limit
If you want yourpost-install/post-upgradechart hooksto work, you should addreadiness probesto your first pod and use--waitflag.helm upgrade --install -n test --wait mychart .pod.yamlapiVersion: v1 kind: Pod metadata: name: readiness-exec labels: test: readiness spec: containers: - name: readiness image:...
So I have a helm chart that deploys a pod, so the next task is to create another pod once the first pod is running.So I created a simple pod.yaml in chart/templates which creates a simple pod-b, so next step to only create pod-b after pod-a is running.So was only at helm hooks but don't think they care about pod status...
Serialize creation of Pods in a deployment manifest using Helm charts
I don't know what am I doing differently now but, the xml is being written to the proper place. I suppose it was a path configuration mistake.ShareFollowansweredAug 8, 2019 at 14:38paulinhorochapaulinhorocha43055 silver badges1313 bronze badges1Did you manage to tell sonar to use this XML file? Is it on 'sonar.testExec...
The coverage output of karma sonarqube unit report comes out as html instead of a xmlI am trying to integrate code coverage during my sonar analysis.I have have coverageify in my stack, i don't know if it is interfering with my output from sonarqube-unit-reporter. In my karma options, i have it do output an ut_report.x...
karma sonarqube unit reporter output comes out as HTML
Based on the description, the--forceflag should do the trick.--force force resource updates through a replacement strategyHowever, there are some issues with it as mentioned in thisGitHub issue.
I have a problem where we essentially discovered a piece of stale configuration in a live environment on one of our deployments (a config map was added as a volume mount). Reading through the docshere(search for 'Upgrades where live state has changed') we can see that helm v2 would purge changes that were introduced to...
How to helm upgrade with v3 and remove / overwrite any manual changes that have been applied to templates
HEAD~0is your latest commit (aka simplyHEAD)HEAD~2represents the hash of the second commit counting from zero.So, typinggit revert HEAD~2you are trying to revert Commit1. That's the difference.
I have 3 commits pushed to my repository.Commit3Commit2Commit1So, if I try to revertCommit2with the commandgit revert commit2Hashit will give an alert in order to solve conflicts before merge.But if I try to revertCommit2with the commandgit revert HEAD~1it will revert Commit2 directly without give me any conflict.Pleas...
What's the difference between revert <hash> and revert <head>?
ngx.thread.spawn not working, only this code worked:access_by_lua ' local socket = require "socket" local conn = socket.tcp() conn:connect("10.10.1.1", 2015) conn:send("GET /lua_async HTTP/1.1\\n\\n") conn:close() ';
How I can duplicate (or create and send) a request with the nginx web server. I can't usepost_action, because it is a synchronous method. Also, I compiled nginx with Lua support, but if I try to usehttp.requestwithngx.thread.spawnorcoroutine, I find the request has been executed synchronously. How do I solve this?locat...
Asynchronous duplication request with nginx
Rather than doing an HTTP proxy, I would use Nginx'sbuilt-in capacityto communicate with uWSGI. (This will still work if you are using separate Docker containers for Nginx and uWSGI since the communication is done over TCP)A typical configuration (mine) looks like this:location / { uwsgi_pass http://127.0.0.1:800...
I've been working on a django app recently and it is finally ready to get deployed to a qa and production environment. Everything worked perfectly locally, but since adding the complexity of the real world deployment I've had a few issues.First my tech stack is a bit complicated. For deployments I am using aws for ever...
Django CSRF Error Casused by Nginx X-Forwarded-host
NOTE: This answer uses boto. See the other answer that uses boto3, which is newer. Try this... import boto import boto.s3 import sys from boto.s3.key import Key AWS_ACCESS_KEY_ID = '' AWS_SECRET_ACCESS_KEY = '' bucket_name = AWS_ACCESS_KEY_ID.lower() + '-dump' conn = boto.connect_s3(AWS_ACCESS_KEY_ID, AWS_SE...
I want to copy a file in s3 bucket using python. Ex : I have bucket name = test. And in the bucket, I have 2 folders name "dump" & "input". Now I want to copy a file from local directory to S3 "dump" folder using python... Can anyone help me?
How to upload a file to directory in S3 bucket using boto
I solved the Problem. It was a plain beginner mistake:- namespaceSelector: matchLabels: namespace: kube-systemI didn't add theLabelnamespace: kube-systemto theNamespacekube-system.After adding the Label it worked instantly.
we are using Rancher to setup clusters with Canal as the CNI. We decided to use Traefik as an Ingress Controller and wanted to create a NetworkPolicy. We disabled ProjectIsolation and Traefik is running in the System project in the kube-system namespace.I created this Policy:# deny all ingress traffic kind: NetworkPoli...
Kubernetes/Rancher: NetworkPolicy with Traefik
OK, I changed the hoster. Now everything works well. I do not know what the problem was exactly. But my hoster did not care about that and I decided to switch to another one.
I am trying to set up a .htaccess file with the following content with Apache 2.2.31:Header always set Strict-Transport-Security "max-age=31536000; includeSubDomains; preload"This is working fine but not for PHP files. The Header is sent twice. I created anemptyHTML file and anemptyPHP file. For the HTML file the heade...
Apache: Created Headers in .htaccess file sent twice when requesting PHP file
You can simply put asonar.propertiesfile under/opt/docker-sonar/conf/. This file will be available inside the container under/opt/sonarqune/conf/, because the folder gets mounted as volume.A full example for asonar.propertiesfile can be found ongithub. However all you need to enter is:sonar.ce.javaOpts=-Xmx<XMX_VALUE -...
I have a Sonarqube instance running as a docker container. Since I updated it to version 7.1 the analysis of my greatest project fails withGC limit exceeded. If I restart the server, it might succeed once. After a while of researching this issue, I am tempted to believe, I need to increase theXmxvalue for the backgroun...
How to change Xmx settings for sonar runner?
No, a soft reset is not enough. Doing this will leave the file in your index (where you stage files to be committed). This means, that git is still tracking the file. You will want to do a mixed reset, which unstages these files as well. As René pointed out, it is also a good idea to remove the file or add it to your ...
I accidentally commited something that might be sensetive information in git (only locally) and I wanna remove it from git history in a simple way. Will git reset --soft HEAD~1 and then unstage the sensetive information and add to gitignore be enough to completely remove it from git history?
does git reset delete history?
It is possible, butit is not possibleto customise the error message.Depending on your function use either:callback("Unauthorized", null);orthrow new Error('Unauthorized');Both of these will produce a 401 response.Seehttps://github.com/awslabs/aws-apigateway-lambda-authorizer-blueprints/blob/master/blueprints/nodejs/ind...
We have our API behind the AWS HTTP API gateway with a custom Lambda authorizer. Our JWT token contains an expiration time and base on that we have to return 401 when it is expired to tell the client to use his refresh token to update JWT.Lambda authorizer returns only 403 even if the token is present but it is expired...
AWS HTTP Api Gateway lambda authorizer how to return 401 if a token is expired
2 Sounds like it is the Console panel blowing up. Consider limiting its buffer size. EDIT: It's in Preferences. Search for Console. Share Follow edited Dec 4, 2009 at 23:21 answered...
We have a process that outputs the contents of a large XML file to System.out. When this output is pretty printed (ie: multiple lines) everything works. But when it's on one line Eclipse crashes with an OutOfMemory error. Any ideas how to prevent this?
How to best output large single line XML file (with Java/Eclipse)?
This is no longer supported as of 4.0.End of Support of WAR deployment ModeThe standalone mode is now the only mode that is supported. Standalone mode embeds a Tomcat server.http://docs.sonarqube.org/display/SONAR/Release+4.0+Upgrade+NotesShareFolloweditedMar 4, 2015 at 16:08schnatterer7,64977 gold badges6363 silver ba...
How can I run Sonar on my Unix system with Tomcat. In previous versions there was way to make .war and deploy it on Tomcat.I tried to put into folder webaps (Tomcat) and run scriptsonarqube-4.1\bin\solaris-x86-32\sonar.sh. Everything was OK, but I didn't know what to write in webbrowser to get to Sonar.Version of my OS...
How to run Sonar 4.1 on Tomcat
0 As of now if you want to schedule a regular bare metal backup of Azure VMs, you can use the Temporary drive to store the files created by WSB if the OS disk size is less than Temporary drive(D:). Else, you have to attach a additional Data disk for backup purpose. I agr...
Since Azure agent doesn't support full VM image backups directly through the portal (without shutting down the VM first), I wanted to schedule a regular bare metal backup of my Azure VMs using Windows Server Backup together with the Azure Backup agent. The challenge is to find a temporary place to store the files cre...
Windows Server Backup and Microsoft Azure Backup
AFAIK you can not create an A record at the zone apex, only an AWS-specific Alias type.The Alias can refer to an ELB, S3 website, CloudFront distribution or another Route 53 record set. You have a couple of options:a) put your instance behind an ELB, and create example.com as an Alias record pointing to your ELB. Or,b)...
I set up an EC2 instance with an elastic IP. I registered a domain with Namecheap and transferred my name servers from them to Route 53.I created an A-IP4 record and plugged in my elastic IP address. Didn't work.Then I decided to try creating the A-IP4 using www. It worked.I've tried setting up a pointer from www.mysit...
Connecting to root domain - AWS Route 53 EC2
Try this.systemctl stop apache2.service
I'm trying to update an SSL certificate on digital ocean with the commandcertbot renewBut I get this error:Problem binding to port 80: Could not bind to IPv4 or IPv6.runningnetstat -pluntshows that port 80 is been used by 'docker-proxy'.What can I do to fix this should I stop docker-proxy how do I do that?
Problem binding to port 80: Could not bind to IPv4 or IPv6 with certbot
-1Yes in most cases you need to open port in your router first (check your NAT settings section), and then setup IIS site binding to 7895 portShareFollowansweredMay 25, 2012 at 7:45Alexander V.Alexander V.1,5281414 silver badges1414 bronze badges6Hi thanks now it's nice I open my port but still haven't see webpage from...
folks I want to know how I can open port for iis.I have also tired from firewall to open port but I can't it's seem I am missing some thing actually i have site (example on port) 7895 in my local I can access it by type localhost:7895 in browser or 192.168.1.1:7895 (local ip) but want to open it through over net for ex...
open port for iis
It is a default network created automatically by docker-compose. Read morehere.
I am new to Docker and found one thing I don't understand: I downloaded the imagejwilder/nginx-proxyfrom the Github reponginx-proxy/nginx-proxyand ran it withdocker-compose up. This will bring on a new networknginxproxy_default, which the new container is connected to, although the docker-compose.yml does not have a ne...
jwilder/nginx-proxy: Where is the network configured? (Not in docker-compose.yml)
0 So try to rename WAMP root folder from C:\wamp46\www to C:\Mirror Edit the httpd.conf file and/or the vhosts.conf file for the site wish to change. The Directory directive will let you specify where the files for this site are to be located. For more info on httpd.conf s...
I got Seagate Backup Plus Slim 1TB today. I am planning to do mirror backup of my web projects from pc (C:\wamp46\www) to external drive (E:). The toolkit app created folders on C:\ and E:\ both named "Mirror" as the syncing folder. Tested it and it works well. But Seagate says: The Mirror folders must each be named ...
Mirror backup WAMP folder into external hard drive
Try disabling theView Results Treein the script as it records all results for you to inspect.The jMeter documentation specifically mentions this:18.3.6 View Results TreeView Results Tree MUST NOT BE USED during load test as it consumes a lot of resources (memory and CPU). Use it only for either functional testing o...
first of all i already had a look at several questions which are quite similar. But i wasn't able to find a solution.My script performs a load test it calls several different URLs(GET http) to download the content behind. After 120 requests the memory usage increases up to 2GB and after 500 to 5-6GBI changed already th...
how to configure Jmeter to discard downloaded files?
The answer is 100 for nginx 1.9.9 or earlier and 1000, for nginx 1.9.10 and later. Thekeepalive_requests directive (default is 100/1000) allows you to configure the maximum number of requests done through a single keepalive connection. From documentation link above: Sets the maximum number of requests that can be ser...
HTTP client could send multiple requests in one HTTP 1.1 connection due to Keep alive feature. But is there any limit of that number in protocol? If not, what is the implementation for Nginx about that? Does it have any configuration?
what is the max number of requests Nginx server allow client to send in one HTTP 1.1 connection
It is because your rewrite rules are infinitely looping. Which is due to the factsection/products/(.*)pattern matches original and rewritten URI.You can use this to fix it:Options +FollowSymlinks -Indexes -MultiViews Options -MultiViews RewriteEngine on RewriteBase / RewriteRule ^(section/products)/([\w-]+)$ $1/produc...
I'm modifying the Apache.htaccessfile for rewrite products' URL, so I can go from thisdomain.com/section/products/product.php?url=some-product-nameto thisdomain.com/section/products/some-product-nameHere's themod_rewritecode that I'm using:Options -Indexes Options +FollowSymlinks Options -MultiViews RewriteEngine on Re...
mod_rewrite issue with GET parameter
May be you have internet connection problem.
While I'm trying to push, pull or merge the local to github repo I'm getting some issues. I even tried to clone the new local repo but that is also giving problem. Can anyone please help me in this matter.Executed command result:$ git push error: Failed connect to github.com:443; Connection timed out while accessing ht...
Error: Can't push, pull, merge or clone in github
You can't match against the query string in theRedirectdirective (nor in aRedirectMatch/RewriteRuleeither). You need to use mod_rewrite's%{QUERY_STRING}var:RewriteEngine On RewriteCond %{QUERY_STRING} (^|&)id=([0-9]+)($|&) RewriteRule ^/?page\.php$ http://test.com/profile/info/id/%2? [L,R=301]ShareFolloweditedSep 10, 2...
I'm trying to figure out a regular expression that will find some numbers in a URL and use them in the URL I redirect to.Redirect 301 /page.php?id=95485666 http://test.com/profile/info/id/95485666i was thinking maybeRedirect 301 /page.php?id=([0-9]+) http://test.com/profile/info/id/$1but it doesn't seem to workAlso, if...
How can I write a .htaccess redirect to find all numbers?
You can disable the response buffer before you return the file result. Response.BufferOutput = false; return File(fileStream, contentType);
I have a file browser application in MVC4 that allows you to download a selected file from a controller. Currently, the FileResult returns the Stream of the file, along with the other response headers. While this works fine for smaller files, files that are larger generate an OutOfMemoryException. What I'd like to do...
ASP.NET MVC: Returning large amounts of data from FileResult
In your Dockerfile you have specified the CMD as CMD [ "/home/benchmarking-programming-languages/benchmark.sh -v" ] This uses the JSON syntax of the CMD instruction, i.e. is an array of strings where the first string is the executable and each following string is a parameter to that executable. Since you only have a ...
I was able to successfully build a Docker image, via docker build -t foo/bar .. Here is its Dockerfile: FROM ubuntu:20.04 COPY benchmark.sh /home/benchmarking-programming-languages/benchmark.sh CMD [ "/home/benchmarking-programming-languages/benchmark.sh -v" ] And here is the file benchmark.sh: #!/usr/bin/env bash #...
Docker gives 'no such file or directory: unknown' on a docker run command
This issue seems to be due to the fact, that opendkim does not set the pseudo resource recordOPT UDPSize, indicating that it can handle responses longer than 512 bytes, as defined byEDNS (wiki),RFC 2671.Opendkim (no EDNS)As can be seen in this tcpdump of an opendkim request:28112+ TXT? selector1._domainkey.outlook.com....
I'm using postfix with opendkim and see a lot of the following errors:opendkim[63]: 84D4C390048: key retrieval failed (s=selector1, d=hotmail.com): 'selector1._domainkey.hotmail.com' reply truncatedThe error occurs for a lot of different domains, but always if a long dkim key (> 1024 bit) is used. I would assume this t...
Opendkim error "key retrieval failed" when long dkim keys are used
Use this one: location = / { index index.html; } location = /index.html { root /your/root/here; }
I would like to nginx to serve a static file from website root ( : http://localhost:8080/ ) but it serves my proxy pass; it serves "/" rule instead of "= /". Here is what my nginx config look like : listen 0.0.0.0:8080; server_name localhost; set $static_dir /path/to/static/ location = / { # got index.html in /pat...
nginx rule to serve root
You can't use the ~ character in Java to represent the current home directory, so change to a fully qualified path, e.g.:file:///home/user1/hbaseBut i think you're going to run into problems in a fully distributed environment as the distcp command runs a map reduce job, so the destination path will be interpreted as l...
I'm trying to back up a directory from hdfs to a local directory. I have a hadoop/hbase cluster running on ec2. I managed to do what I want running in pseudo-distributed on my local machine but now I'm fully distributed the same steps are failing. Here is what worked for pseudo-distributedhadoop distcp hdfs://localhost...
Backup hdfs directory from full-distributed to a local directory?
if directive works before your request send to backend, so at that time there is no $sent_http_... variable. You could use map directive. log_format main_log '$extended_info'; map $sent_http_x_extended_info $extended_info { default $sent_http_x_extended_info; "" "-"; }
I'm trying to log data from custom header. In response: Cache-Control:no-cache Connection:keep-alive Content-Type:application/json Date:Mon, 09 Nov 2015 16:09:09 GMT Server:nginx/1.9.4 Transfer-Encoding:chunked X-Extended-Info:{"c":70} X-Powered-By:PHP/5.6.12 In php script (Symfony2): $response->headers->set('X-Exten...
nginx read custom response header
I'm using GitKraken version 4.0.5 (MacOS and Windows) and spaces are shown:You have however a designated button for hiding spaces (on the top-right corner):Which results (same source with spaces):Maybe is it turned on on your client?
I use GitKraken and and it is really cool tool! Is it possible to see spaces in GitKraken?For example, there are spaces, but GitKraken shows no spaces:But another git visual tool shows spaces:Is it possible to see spaces in free version of GitKraken?
How to see spaces in GitKraken free version
You can set the pointers to NULL, then the destructor will not delete them.struct WithPointers { int* ptr1; int* ptr2; WithPointers(): ptr1(NULL), ptr2(NULL) {} ~WithPointers() { delete ptr1; delete ptr2; } } ... WithPointers* object1 = new WithPointers; WithPointers* object2 ...
I have an object with some pointers inside of it. The destructor callsdeleteon these pointers. But sometimes I want to delete them, and sometimes I don't. So I'd like to be able to delete the objectwithoutcalling the destructor. Is this possible?Edit: I realize this is an AWFUL idea that no one should ever do. Non...
Is it possible to delete an object in c++ without calling the destructor?
On current MacOSes you would want to use packet filtering:https://blog.neilsabol.site/post/quickly-easily-adding-pf-packet-filter-firewall-rules-macos-osx/
I am connected to a network and there is a particular node that keeps scanning me and attempting to connect to me. It is always from the same IP. I have looked and cant seem to find a way to block that IP on my MAC. Is there a way to drop this particular IP on my MAC?
block IP address on MAC
The images are probably being cached. Take a look at[img setCacheMode:]Did you actually try doing 500 times or are you guessing at the behaviour? My guess would be that the cache would be cleared at some upper limit - maybe 50mb is not that much?It is important to note that-releaseis not equivalent tofree()ordestroy(),...
NSImage *randomImage = [[NSImage alloc] initWithContentsOfURL:imageURL]; [randomImage release];Why does the memory usage still go up? What is using that memory? I release the NSImage object. ( no, its not the URL )
NSImage + memory managament
-1As far as I know, when the private repo is yours and someone else opens a pull request to it, you as the owner have to merge the pull request in the respective branch.
I have a protected github repository, where I want a user that was already allowed 'read' access to also be able to merge PR's, so I gave him the 'write' role. According to thegithub docsthat should be enough. Still he is not able to merge, and he sees a warning about not having write access. Am I missing something?
Github organization member can't merge PR's even though he has write access
No, this is not supported, you might be able to hack your way through, but certainly not out of the box.But you can create an internal load balancer for your service in the network and its ip wouldnt change, you do this using a service with an annotation:--- apiVersion: v1 kind: Service metadata: name: name annotat...
I am having an aks instance running. which I assigned an virtual network to it. So all the Node IPs in the network are good and I can reach them from within the network.Now I wonder if it is possible to create a 2nd virtual network and tell kubernetes to use it to assign public ips ?Or maybe is it possible to say that ...
assign kubernetes loadbalancer an ip from an internal network
I found the answer to my question here:http://www.mos-eisley.dk/display/it/Elasticsearch+Dashbord+in+Grafana. You can ignore the parts about setting fieldname=true and instead just set it to query the fieldname.keyword when creating the template.Just a quick note: Something that took me too long to realise is that when...
I have an Elasticsearch (5.1.2) data source and am visualizing the data in Kibana and Grafana (4.1.1). For string values in my dataset I am using the keyword feature as described athttps://www.elastic.co/guide/en/elasticsearch/reference/5.2/fielddata.html. An example of the mapping for fieldname "CATEGORY":"CATEGORY": ...
Grafana cannot aggregate on String fields as it does not recognize keyword field in Elasticsearch
The case you're describing is not what Git typically considers a rename. Generally, a rename in Git is when one file is removed in the same commit as another file is added and the files are identical or similar. In your case, the old file hasn't been removed, so you now have two files. If they are identical or simil...
Some time ago, someone pushed a file to the github repository whose name is the wrong case. Now there is a new version of the file, which has the correct case in its filename. When I push the new version to github and create a pull request, the "Files Changed" view shows a new file and no changes to the old file. In o...
How to change the case of a filename
Use awk command instead docker images | grep "none" | awk '{print $3}' | xargs docker rmi
I'm trying to delete each docker image that it's name == none in my system. I have tried this for image in $(docker images | grep none); do echo $image; done But this gives me the output of each column like that: <none> <none> a20d00ca4041 19 minutes ago 227MB I want it li...
docker images output manipulation
Amazon EKS uses IAM to provide authentication to your Kubernetes cluster through the AWS IAM Authenticator for Kubernetes. You may update your config file referring to the following format:apiVersion: v1 clusters: - cluster: server: ${server} certificate-authority-data: ${cert} name: kubernetes contexts: - co...
I have to setup CI in Microsoft Azure Devops to deploy and manage AWS EKS cluster resources. As a first step, found few kubernetes tasks to make a connection to kubernetes cluster (in my case, it is AWS EKS) but in the task "kubectlapply" task in Azure devops, I can only pass the kube config file or Azure subscription ...
How to connect AWS EKS cluster from Azure Devops pipeline - No user credentials found for cluster in KubeConfig content
Navigate to Setting->Emails (or directly tohttps://github.com/settings/emails), and just add the emails you used on your "unattributed" commits.
I just finished a semester at college and decided to import all of my projects from bitbucket (required for my classes) to github (where all of my other projects are). I successfully imported them. Unfortunately, at the time when I was working on these projects, I was switching between three different computers.As a re...
How can I set a username alias in github commits?
If your goal is to run an Apache Web Server (httpd), you should use thehttpd image.Docker containers are generally meant to run a single process. So, you wouldn't normally design a container to run something like systemd as the root process, and then run httpd as a child process. You would just run httpd directly in th...
I am using archlinux/base official image from docker hub.I am trying to use systemctl and it says.$ docker run --rm -it ac16c4b756ed systemctl start httpd System has not been booted with systemd as init system (PID 1). Can't operate.How to solve this.
docker archlinux image: System has not been booted with systemd as init system (PID 1). Can't operate
8 You can create your own AMI but you need to use the Amazon-supplied kernels. The newest they provide is 2.6.21. I have a list of the fc (Fedora Core) kernels that I use for CentOS instances. I'm pretty sure they work fine with Ubuntu as well. You'll want to bake these i...
I have an Amazon EC2 instance using the Amazon-supplied Fedora 8 64-bit AMI, which I would like to upgrade to Fedora 10. I tried doing this by running "yum update" to upgrade the kernel and all packages. This seemed to work fine and I see that I now have the fc10 kernel installed, and all of my installed packages hav...
How does an Amazon EC2 instance select its kernel?
2 AWS Lambda comes with an ephemeral storage unit in /tmp. However, please note that the ephemeral storage unit still has a storage of 512MB. You can load your dependencies to this storage, and write code accordingly. Share Improve this answer ...
I am new to AWS Lambda and running a tensorflow model in AWS Lambda. Now tensorflow 1.0.0 is the one that fits into the 50Mb limit but since tensorflow 2.0 is much bigger in size it does not fit. Does anyone knows of a way to use tensorflow 2.0 with AWS lambda?
How to use tensorflow 2.0 with AWS Lambda?
You can switch it off in the server configuration.The parameter name isenable_result_cache_for_sessionas mentioned inthe Redshift documentation.
Since the 21 November Amazon Redshift introduced the default caching of result sets. Is there a way to disable caching by default on a Redshift database? There don't seem to be many docs on it at the moment.
disable caching on redshift by default
Within your foreach loop, you are using array_push. You are adding to the array you are iterating through, this is an infinite loop.
Im trying to build an array that needs to be passed to a report. Some of the data returned has similar field names so im using the function below to add a prefix to the array key names before merging the arrays, however i get an out of memory exception "Fatal error: Allowed memory size of 536870912 bytes exhausted (tr...
Fatal error: Allowed memory size of 536870912 bytes exhausted
This is because/teammaps to an existent directory. When you request/teamserver changes the uri to/team/adding a directory slash thus it goes to the dir.You have to turn off the DirectorySlash.Add the following line to your .htaccessDirectorySlash offThis will allow you to access/team.phpas/teamwithout trailing slash.Yo...
How can I set up an htaccess that can distinguish a file from a folder with the same name?I have under my websiteindex.php team.php team/Justin.php team/martin.php...and a htaccess with a URL Rewrite to make nice url and remove the .phpRewriteEngine On RewriteCond %{REQUEST_FILENAME} !-f RewriteRule ^([^\.]+)$ $1.php ...
htaccess identical folder and file name
dataList = [dataArr];this is not valid Objecitve-C. If you wanted to writedataList = dataArr;that's still a no-go, as you're acessing the instance variable directly, not through the property setter, that is, your array won't be retained and it will badly crash.[dataList release]; [dataArr retain]; dataList = dataArr;is...
My purpose: making an API call to a server, and getting back from them is an array of data nameddataArrand I want to store these data to another array for later need.What I am doing so far ismyClass.h:@propery ( nonatomic, retain ) NSArray *dataList;myClass.m:@implementation myClass -(void)receivedData:(NSArray*) dataA...
save data to another array, memory management, Objective C
The response header gives: Content-Type: application/javascript This is the MIME type that needs to be included in your gzip_types statement in order to compress these types of response. Your existing value contains many similar MIME types, but not one of them is an exact match for what the server actually sends. See...
This is our Nginx configuration with regards to Gzip: gzip on; gzip_disable "msie6"; gzip_vary on; gzip_proxied any; gzip_comp_level 5; gzip_min_length 256; gzip_buffers 16 8k; gzip_http_version 1.0; gzip_types text/plain text/css application/json...
Nginx - Amazon Cloudfront - Gzip Doesn't work for JS files
Other answers suggest the data might be stored in: C:\Users\Public\Documents\Hyper-V\Virtual hard disks\MobyLinuxVM.vhdx or since the Windows 10 Anniversary Update: C:\ProgramData\docker\containers You can find out by entering: docker info Credit to / More info: https://stackoverflow.com/a/38419398/331637 https...
I'm currently experimenting with Docker containers on Windows Server. I've created a number of containers, and I want to see where they are actually saved on the host's file system (like a .vhd file for Hyper-V). Is there a default location I can look, or a way to find that out using Docker CLI?
Where are containers located in the host's file system?
0 Your expectation is incorrect. --oom-kill-disable does not disable virtual memory overcommitment, which would cause malloc to fail if mmap fails to allocate requested pages. Instead, the option maps to the cgroup v1 feature exposed via sysfs (/sys/fs/cgroup/memory/docker/...
I wrote a short java program to allocate memory: package com.company; import java.util.ArrayList; import java.util.List; public class Main { public static final int SIZE_NATIVE_LONG_IN_BYTE = 8; public static void main(String[] args) { Integer memoryConsumptionInMiB = Integer.parseInt(args[0]); ...
docker oomkiller vs malloc failure from within the container
6 I played around with the USER command in the Dockerfile, but could never get it to work with an admin user. However, I found in a GitHub posting the mention of specifying the user in the docker run command like this did: docker run --user "NT Authority\System" ... Which...
I have a .NET Core app that's required to be "run as administrator" and I'm trying to get it to be built into a Docker image. I am able to build a Docker image just fine, but it fails at runtime with the "Need to run as Administrator" error. Is there a way in the Dockerfile or in the docker run command to specify this...
Can't get a Windows Docker container to "run as administrator"
You can filter out metrics based on the other metrics withunlessoperator. It removes metrics from left-hand-side of this operator with same values of labels as those at the right-hand-side.For example if you have metricsmetric1{label1="value1"} metric1{label1="value2"} metric2{label1="value1"}expressionmetric1 unless ...
Let's say I have the following metrics:system_cpu_usage{hostname="host1"} 10 system_cpu_usage{hostname="host2"} 92 system_cpu_usage{hostname="host3"} 95 process_cpu_usage{hostname="host2", cpu_usage="high"} 90I have an alert condition as follows:avg_over_time(system_cpu_usage[5m]) > 90Which returns all instances where...
Filter Prometheus metrics by label of another metric
As git isdecentralized(which actually meansdistributed), you can safely copy your original clone entire folder (the one with the.gitsub-directory) from your old machine to the new one without losing anything. Even your local commits or branches will be kept as git embeds the complete distant AND local history in each c...
I have seen some information on this topic, but didn't see a definitive agreement on the proper strategy.I just got a new macbook and have been setting everything up over the past few days. I have a git repo on the old machine that I want to move over to the new machine, what is the best way for me to move the entire ...
Moving local repo to a new Macbook
6 No. Support for Zenodo is an open issue. Contact was made with Zenodo asking for Bitbucket support, but Zenodo said Bitbucket currently lacks the right features in the API. Share Follow answered ...
I see that GitHub has DOI integration through Zenodo but is there an equivalent tool for Bitbucket? Or do I have to contact a DOI Registration Agency directly?
How to create digital object identifier (DOI) for bitbucket repository?
In your last block try making theprintline come beforeresult[dnsIpAddress] = "FAILURE"My guess is either there is more code than what is shown here or the line before the print statement causes a different exception.
While trying to implement aDNSRequest, I also needed to do some exception handling and noticed something weird. The following code is able to catch DNS requesttimeoutsdef lambda_handler(event, context): hostname = "google.de" dnsIpAddresses = event['dnsIpAddresses'] dnsResolver = dns.resolver.Resolver() ...
Exception handling in aws-lambda functions
This will probably be the case of your result data set exceeding the limit 1MB:If the total number of scanned items exceeds the maximum data set size limit of 1 MB, the scan stops and results are returned to the user as a LastEvaluatedKey value to continue the scan in a subsequent operation. The results also incl...
In a DynamoDB table, I have an item with the following scheme:{ id: 427, type: 'page', ...other_data }When querying on primary index (id), I get the item returned as expected.With ascanoperation inside AWS DynamoDB web app to get all items with typepage, 188 items including this missing item are returned. H...
DynamoDB scan leaves valid item out
Actually i figured out this CAN be done via the api in this way, it just requires headers and data indicating what permissions.: curl -H "Accept: application/vnd.github.v3+json" -u YourUserName:YourPersonalAccessToken -X PUT -d '{"permission":"write"}' https://api.github.com/teams/$team_id/repos/$org_name/$repo Alter...
It is possible to add collaborators via the api as described here: https://developer.github.com/v3/repos/collaborators/#add-user-as-a-collaborator Endpoint: /repos/:owner/:repo/collaborators/:username But what about adding team access, which is definitely possible via web interface in "Settings > Collaborators & Teams...
Is there a github api endpoint to give Team access to a repo?
c->data = &data; stores the address of the pointer data (the argument to your function), not the actual pointer. I.e., you're storing a pointer to a temporary. You could have built the container structure with just a void *data member.
I just wrote some C code: #include <stdlib.h> #include <time.h> #include <string.h> typedef struct { void **data; time_t lastModified; } container; container *container_init() { container *c = malloc(sizeof(container)); void *data = NULL; c->data = &data; c->lastModified = time(NULL); ret...
C - memory management
Jenkins itself will happily run on a micro, but there are two problems: 1) you won't have much memory left for building and testing, around 150MB, but the bigger problem is 2) if your CPU usage spikes for more than a few seconds Amazon will simply crush your instance with throttling cutting off 97% or more of availabl...
I am planning to install Hudson on Amazon EC2 using Ubuntu image. The code I am going to test does not have a big memory overhead - I will be executing mainly python unit tests. Which EC2 instance should I use? Would micro instance be sufficient (have enough memory) or should I use a bigger instance?
Running Hudson on EC2
I opened a ticket with AWS support and they were able to find the IP that was consuming the read capacity. They used an internal tool to query logs that are not available to customers. They also confirmed that these events do not get emitted to Cloudtrail logs, which only contain events related to the table, such as r...
We have a DynamoDB table that we thought we'd be able to turn off and delete. We shut down the callers to the web services that queried it (and can see on the web server metrics that the callers have dropped to zero), but the AWS console is still showing Read Capacity consumption greater than zero. However, every oth...
How can I find out what is consuming my DynamoDb tables Read Capacity?
You cannot access the env context in matrix. You can use a job with outputs to set the matrix: env: SERVICES_JSON: | [ "a", "b", "c" ] jobs: gen-matrix: runs-on: ubuntu-latest steps: - name: Generate Matrix id: gen-matrix run: | # use heredoc with ...
I have a list of service which I want to pull, and for that i wanted to use matrix. Since this list of services defined in SERVCES_JSON env variable will also be used in other jobs, I would like to reused in to convert to list, instead of defining the same list again. name: Deployment on: workflow_dispatch: env: ...
matrix in GitHub Actions: how to use a json defined in env variable as matrix list
Browsers will usually get this information through HTTP headers sent with the page.For example, the Last-Modified header tells the browser how old the page is. A browser can send a simple HEAD request to the page to get the last-modified value. If it's newer than what the browser has in cache, then the browser can relo...
This is a dangerously easy thing I feel I should know more about - but I don't, and can't find much around.The question is:How exactly does a browser know a web page has changed?Intuitively I would say that F5 refreshes the cache for a given page, and that cache is used for history navigation only and has an expiration...
How does the browser know a web page has changed?
myteamid = teamidjson(['id']) That seems to be causing the error. The correct way to access the id key is: myteamid = teamidjson['id']
Having difficulty parsing json from GitHub api. I'm trying to populate a team with all the repos from an organisation. I'm using myteamname to obtain the teamid required for the loop which populates the team with the repo names. import json import requests mytokenid = "xxx" myorg = "xxx" myteamname = "xxx" headers ...
Parsing json GitHub api with Python
sbin is not in the path when run via cron. Specify the full path to service. This is probably either /sbin/service or /usr/sbin/service. You can find the path on your system by running which service.
service service_name start When i tried running this from cmd line, it works. But when i try to schedule it via cron, i get an error saying /bin/sh: service: command not found
Unable to run a service command via cron
+50Your compare URL could be something likehttps://github.com/gaganmalvi/kernel_xiaomi_lime/compare/Q..02ca1a9Qis the name of the only branch in that repository. It is used here for the HEAD of the repository.02ca1a9is the "Git Object ID" for the state of the repository after the last commit in Dec 2020.GitHub document...
I'm trying to see all modifications made from 06e27fd143240e8e4d13b29db831bedece2bf2d3 to the latest e1c34175b5556ac5ce1e60ba56db2493dd9f6b52. I triedhttps://github.com/gaganmalvi/kernel_xiaomi_lime/compare/Q:e1c34175b5556ac5ce1e60ba56db2493dd9f6b52%5E%5E%5E%5E%5E...Q:06e27fd143240e8e4d13b29db831bedece2bf2d3and vice-ve...
How to see changes from commit x to y on github?
Try removing thetablepart from your--formatargument, such as:docker ps --format '{{.Names}}'It should give you a simple list of container names with no table heading
docker ps --format "table {{.Names}}"outputNAMESin first row:root@docker-2gb-blr1-01:~# docker ps --format "table {{.Names}}" NAMES enr osticket osticket_db ...docker inspect --format '{{.Name}}' $(docker ps -q)prints/in the beginning of container name:root@docker-2gb-blr1-01:~# docker inspect --format '{{.Name}}' $(do...
'docker ps' output formatting: list only names of running containers
Doing agit pullshould do the right thing, as long as you haven't done git add on the files you don't want in to have under git. I suggest putting the names of those files in a .gitignore.If you are running into a specific problem with usinggit pull, you should ask aboutthat.I don't know much about DaftMonk, but if it ...
How can I pull down a git and have it overwrite my local project ONLY where conflicts are found?E.g. I have many folders / files in my local project that are not on the git project and never will be.Ok... here is the full scenario.I used DaftMonk generator to create a fullstack boilerplate:https://github.com/DaftMonk/g...
How to pull files and only override conflicts
I think that is not possible because :PVC is a namespaced resource and PV is not a namespaced resource.kubectl api-resources | grep 'pv\|pvc\|NAME' NAME SHORTNAMES APIVERSION NAMESPACED KIND persistentvolumeclaims pvc v1 ...
Below is my scenario. I have an NFS setup and it will be used to create PV. and then use PVC to bind the volume. Now, Consider I want to bind particular PV/PVC together irrespective of where PVC will be created. As far as I tried I could not bind PV/PVC without bringing namespace into the picture. Since I use helm char...
is namespace mandatory while defining claimRef under k8s PersistentVolume manifest file?
"mocha: command not found" means you have to install mocha in your gitlab runner environment.test: stage: test script: - npm install --global mocha - mocha test
I want to try CI/CD. So I am working on a simple project. I wanted to run the test file. But I get the error "mocha: command not found". There is no problem when I try it in my own terminal. How can I solve this?Thanks.
mocha: command not found in GitLab
Posix is a standard, not a specific set of code, but we can look at libc for an example. Here's what posix_memalign() initially allocates in that implementation. mem = malloc (size + 2 * alignment); With this beautiful ASCII illustration. /* ______________________ TOTAL _________________________ / ...
I am trying to decide if I should use memalign() over malloc() because aligned memory would make my job easier. I read the GNU documentation here (http://www.gnu.org/software/libc/manual/html_node/Aligned-Memory-Blocks.html) which mentions that The function memalign works by allocating a somewhat larger block. I want ...
how much extra memory does posix_memalign() take?
12 You can do both deployment and cache invalidation with the help of aws-cli. #!/bin/bash # enable cloudfront cli aws configure set preview.cloudfront true # deploy angular bundles aws s3 sync $LOCAL s3://$S3_BUCKET \ --region=eu-central-1 \ --cache-control max...
we have our Angular2 code in S3 .And we access it via Cloudfront. It works fine. But after a deployment to Angular2 , we want every code to be invalidated from Cloudfront. What are the best approaches for clearing cache after deployment? How to handle cloudfront caching?
How to handle cloudfront cache after deployment
Developers can not add to this feature. Microsoft scanned over 100,000 GitHub repositories and looked at popular repositories to get these examples which I read awhile back inIntelliCode with API usage examples
In Visual Studio, when you hover overSystem.Reflection.MethodInfo.GetCustomAttributes(see definition), it has a link at the bottom "GitHub Examples and Documentation". When you click on that link, it opens these examples directly in Visual Studio.Does anyone know how this is implemented in the XML Code docs? Because th...
How can I put GitHub examples in XML code docs?
As mentioned in the other answers, the list() call is running you out of memory. Instead, first iterate over maxcoorlist in order to find out its length. Then create random numbers in the range [0, length) and add them to an index set until the length of the index set is 1000. Then iterate through maxcoorlist again an...
OK, so I have a problem that I really need help with. My program reads values from a pdb file and stores those values in (array = []) I then take every combination of 4 from this arrangement of stored values and store this in a list called maxcoorlist. Because the list of combinations is such a large number, to speed ...
Python Memory Error when using random.sample()
Backups are per-device. So a backup of your iPod will not be restored to your iPhone. In other words, there is no sync.
When does data get restored for an app? What if I save data in the app's document directory. Then they sync with iTunes. Now iTunes has a backup. Will that data be populated to another device when they sync that new device to their iTunes or will they just get a clean install of my app? I'm trying to figure out how to...
iPhone when does data get restored from backup
There are pros and cons to using lambda functions as your AppsSync resolvers (although note you'll still need to invoke your lambdas from VTLs): Pros Easier to write and maintain More powerful for marshalling and validating requests and responses Common functionality can be more DRY than possible with VTLs (macros ar...
I have been looking into AWS AppSync to create a managed GraphQL API with DynamoDB as the datastore. I know AppSync can use Apache Velocity Template Language as a resolver to fetch data from dynamoDB. However, that means I have to introduce an extra language to the programming stack, so I would prefer to write the res...
AWS AppSync Resolvers Lambda Function vs Velocity Template Language (VTL)
You don't need to refork again. Just add a remote (say, upstream) and fetch upstream to update your cloned repository. $ git remote add upstream <original-repo-url> $ git fetch upstream # update local with upstream $ git diff HEAD..upstream/master # see diffs between local and upstream/master (if ...
I created the fork of some GitHub project. Then I created new branch and did a patch inside of that branch. I sent the pull request to author and he applied my patch and added some commits later. How can I synchronize my fork on GitHub with original project now? Am I to delete my fork on GitHub and create new fork for...
How to synchronize fork with original GitHub project?
There are numerous ways to do this; in the end, we went forchanging the permissions (READ/WRITE/ADMIN) on (team, repository) combinations via the REST API.That's not to say that webhooks, enabling/disabling branch restrictions, or the pre-merge would not work, however.
Here, we use GitHub Enterprise. We have an issue with people accidentally merging PRs during code freeze windows, which interferes with our in-house release tool. It would be nice if we could find a way to prevent this.What I'm trying to do, is find a way to disable the big green Merge button on each repo belonging to...
GitHub Enterprise: enforce code freeze during release?
Based on your use case you can utilized service discovery feature of ECS, service discovery will give an endpoint(url) to communicate between different services privately. In service discovery ECS take care of updating dynamic IP and port of containers to DNS record, every time a new task is started or stopped. Refere...
I didn't find a solution for that two containers in separate task definitions can communicate with each other. Therefore, I follow the answer to link the two containers in the same task definitions which works well.Thanks for the answer first. However, when I read the ECS documentation, I find the following paragraph ...
How to make containers communicate with each other in ECS without link and port mapping?
You can use directives like this to allow an IP range for certain URL:# set env variable if URL is /rest or /rest/ SetEnvIf Request_URI "/rest(/.*)?$" rest_uri Order deny,allow # first deny all Deny from all # then allow if env var is not set Allow from env=!rest_uri # also allow your IP range Allow from 10.1.0.0/1...
I would like to block a path from my site using the .htaccess configuration. The idea is that only a specific set of IP's can access that specific path from the URL.Note:It's a path, not a page or directory. We are trying to shield off a web-service so there will be only post calls to the URL's.I would like the urlexam...
.htaccess path only accessible by ip
I had the same problem due to restclient misconfiguration. Have a look how restclient is created and configured in the examplehere.
I have made a kubernetes operator using this frameworkhttps://github.com/operator-framework/operator-sdkin which I have a small custom resource definition defined and a clientset generated.I create a client for this custom resource doing:imports are ( "k8s.io/client-go/kubernetes" "k8s.io/client-go/rest" ) con...
encoding is not allowed for this codec: *versioning.codec
Hi you can active "required string" functionality of web monitoring. It use regular expression pattern. Ciao!Required string: Required regular expression pattern. Unless retrieved content (HTML) matches the required pattern the step will fail. If empty, no check on required string is performed. For example: Homepage of...
I am using Zabbix 5.4.3 to monitor all of my company hosts.I want to monitor a local website address (eg.https://172.30.200.1:44443/login) which is our firewall webpage.It has got two linked WANs, one with our primary public IP and another which is a 4G backup connection without public IP (random access IP).When the co...
Zabbix 5.4.3 - How to monitor a string in a webpage and define a trigger when it changes
Its not possible to filter objects by regular expression. It is possible to filer object by lableThis is the code that will filter by labellabelSelector := labels.Set(map[string]string{"mylabel": "ourdaomain1"}).AsSelector() informer := cache.NewSharedIndexInformer( &cache.ListWatch{ ListFunc: func(options...
Im writing custom controller for kubernetes. Im creating shared informercache.NewSharedIndexInformer( &cache.ListWatch{ ListFunc: func(options meta_v1.ListOptions) (k8sruntime.Object, error) { return client.CoreV1().ConfigMaps(nameSpace).List(options) }, ...
kubernetes filter objects in Informer
0 is your desired output is something like: $host/mnt/synology/Torrents/Games/ where $host is the name of each one of these ips: (192.168.1.40 192.168.1.41 192.168.1.42 192.168.1.43) ? when building the path for mkdir you are doing $(hostname) but that command's output wil...
I'm trying to learn to write some simple bash scripts and I want to create a backup script that will use rsync to fetch predetermined directories and sync them to a backup machine. Here is the code: #!/bin/bash #Specify the hosts ip=(192.168.1.40 192.168.1.41 192.168.1.42 192.168.1.43) #currently unused webdirs=(/et...
How do I loop through an array of ip adresses to get the hostname of each machine in bash?
Your image doesn't have a command calledecho.AFROM scratchimage contains absolutely nothing at all. No shells, no libraries, no system programs, nothing. The two most common uses for it are to build a base image from a tar file or to build an extremely minimal image from a statically-linked binary; both are somewhat ...
I have a docker image with the following dockerfile code:FROM scratch RUN echo "Hello World - Dockerfile"And I build my image in a powershell prompt like this:docker build -t imagename .Here is what I do when I build my image :Sending build context to Docker daemon 194.5MB Step 1/2 : FROM scratch ---> Step 2/2 : RUN...
Docker: Run echo command don't work on my window container
It looks like what might be going on is the default.conf file in the nginx image is taking over the / location. You nginx run command has: -v $ROOT/web/flask/conf/nginx-default.conf:/etc/nginx/conf.d/default \ This should be overwriting the default.conf instead of just default. As it currently stands, it just adds an...
I am trying to connect docker nginx with docker flask. Here is the structure of my project: . ├── storage │   ├── nginx │   │   └── static │   │   └── image.gif └── web └── flask ├── app │   ├── run.py │   └── templates │   └── index.html ├── conf │   ├── ngi...
nginx error 403 - directory index is forbidden
You have one script which does that by: getting the last commit of each branch checking that commit is part of the history of master That would delete rebased branches which have been merged to master. last_commit_msg="$(git log --oneline --format=%f -1 $branch)" if [[ "$(git log --oneline --format=%f | grep $last_c...
in our team we keep the fast-forward only merge policy for master and development branches in order to prevent merge commit hell: I do not delete my topic branches once they are merged (or rebased and then merged), so I end up with tons of these. I can delete some: git branch --merged This will only show me those wh...
Mass deleting local branches that has been rebased and merged
If it is a public cluster where each node in the cluster has an ip address the public ip will be the address of the node the pod is on. If it is a private cluster you can deploy a nat gateway for all the nodes and specify static ip addresses.you can use this terraform module for a private cluster:https://github.com/te...
Pod A is on ClusterIP service type, so incoming requests from external resources are not allowed. Pod A executes outgoing requests to 3rd party services (Such as Google APIs). And I want to specify the IP address that this request is coming from on google for security reasons.Is there a way to find the IP address this ...
Kubernetes Pod ipv4 address for outgoing http request
1 I figured out, request was completely incorrect: This one gonna work out: https://api.github.com/search/repositories?q=goit-js+user:realtril&per_page=1000 Share Improve this answer Follow answe...
What I wanna do is just get the same filtering result as I am getting in github.com: As you can see it's 13. But when I am doing the request like that: const ghReq = await fetch( 'https://api.github.com/users/realtril/repos?q=goit-js&per_page=100' ); const ghData = await ghReq.json(); console.log(ghData); I am get...
Filtering by name via GitHub API is not giving the correct result
Use the reset subcommand:git checkout A git reset --hard B git push --force githubAs a sidenote, you should be careful when usinggit resetwhile a branch has been pushed elsewhere already. This may cause trouble to those who have already checked out your changes.ShareFollowansweredMay 27, 2010 at 16:56Bram SchoenmakersB...
The title is not very clear. What I actually need to do often is the following:Let's say I have a development going on with several commits c1,c2,... and 3 branches A,B,Cc1--c2--c3--(B)--c4--(A,C)Branch A and C are at the same commit.Now I want branch A to go back where B is, so that it looks like this:c1--c2--c3--(A,B...
How to move a branch backwards in git?
The problem had to do withkubeadmnot installing a networking CNI-compatible solution out of the box;Therefore, without this step thekubernetesnodes/master are unable to establish any form of communication;The following task addressed the issue:- name: kubernetes.yml --> Install Flannel shell: kubectl -n kube-system a...
I have set up my master node and I am trying to join a worker node as follows:kubeadm join 192.168.30.1:6443 --token 3czfua.os565d6l3ggpagw7 --discovery-token-ca-cert-hash sha256:3a94ce61080c71d319dbfe3ce69b555027bfe20f4dbe21a9779fd902421b1a63However the command hangs forever in the following state:[preflight] Running ...
Joining cluster takes forever
If each of those String Arrays are big "enough" and it appears you do want to store them - have you considered Sqlite? SharedPreferences is most effective to store primitive data in key-value pairs. Check this link - it has neat comparison about the options you have - http://developer.android.com/guide/topics/data/da...
In my app I have 5 String arrays that represent different fields of objects. i.e. String_A[1], String_B[1], String_C[1], String_D[1], String_E[1], All are attributes of the same object (which is not really an object). Now I want to store those in order to be able to use them in a new activity that I am creating. Si...
Most effective way of storing Strings in Android
Here is the information you're looking for :https://docs.aws.amazon.com/aws-backup/latest/devguide/s3-backups.html#S3-backup-limitationsBackup size limitations: AWS Backup for Amazon S3 allows you to automatically backup and restore S3 buckets up to 1 PB in size and containing fewer than 24 million objects.
I am using AWS Backup to back up S3 buckets. One of the buckets is about 190GB (the biggest of the buckets I am trying to back up) and it is the only bucket that the backup job fails on, with the error message:Bucket [Bucket name] is too large, please contact AWS for support The backup job failed to create a recovery ...
AWS Backup for S3 buckets - what is the size limit?