Your adult wordpress on autopilot – i will schedule image posts for 2 months for $10

Your adult wordpress on autopilot – i will schedule image posts for 2 months

Hello. I can make your adult wordpress run on autopilot for 2 months for just $10.

  • I will schedule image posts (around 3-5 posts/day).
  • I can scrape images myself or you can provide your photos.
  • Just give me your niche and i will do the rest.
  • I will write general titles that will fit into any picture
  • I can write general descriptions that will fit into any picture, i can include links that you provide

All images will be hotlinked from so you won’t lose any bandwitch, but if you want i can upload images to your wordpress.
If you need longer schedule you can multiple order, or just buy 1 additional month for $5


How can I schedule a Google Calendar event for recurring twice-monthly payday?

I found the answer at

Create a file called payday_mid-month.ical containing:

SUMMARY: Deposit Paycheck

Create a file called payday_end-of-month.ical containing:

SUMMARY: Deposit Paycheck

Import each of them at

Contrary to what others had reported, I was then able to edit those recurring events in Google Calendar, and my edits flowed through to all instances of that event. (I was able to edit the title, description, calendar, time of day, notifications, etc, as long as I was willing to leave the time zone as UTC.)

cron – How can you schedule regular tasks to execute when available?

I want to schedule some tasks on my ubuntu 20.04 destkop to execute regularly. Like backups or downloads from certain sites or whatever.

I have been using crontab for this, but if so happens that my computer is not powered at that time, it will just not do them.

How can I schedule a task to run, lets say, every saturday at 12.00pm or whenever the computer is back online after that date?

reward schedule – The End of Mining

I would like to add some other effects not mentioned yet. The income of newly mined bitcoin to miners will get less over time. The costs don’t outweigh the benefits, so miners drop out. This results in longer processing time per block. Then it would cause the difficulty to drop, making it less expensive to mine a block.

So there won’t by a mining stop, there will just be less miners mining against a lower difficulty. At the level where the benefits outweighs the costs. And since the amount of new coins mined is halve every 4 years, the changeover is not abrupt.

Partly this will be compensated by the transaction fee, but people will also be able to change to other crypto-currencies if those provide lower fees. This could effect in a price drop of BTC, which would make it less lucrative for miners to continue, again re-establishing the balance.

Miners don’t hold an absolute monopoly.

In real live economy: The consumption of milk is going down, so there are too many milk farmers. The price of the milk goes down. Some milk farmers won’t be able to feed their cows and will close doors. Some of them do things smarter / cheaper and will survive.

Snapshot schedule is not attaching boot disk

I created a VM and snapshot schedule with terraform modules on GCP. The code is attaching the additional disks but not the boot disk. Any idea what is needed to be changed to the below code to include the boot disk.

Any help will be appreciated.

locals {
attached_disks = {
for disk in var.attached_disks : => merge(disk, {
options = disk.options == null ? var.attached_disk_defaults : disk.options
attached_disks_pairs = {
for pair in setproduct(keys(local.names), keys(local.attached_disks)) :
“${pair(0)}-${pair(1)}” => { disk_name = pair(1), name = pair(0) }
iam_roles = var.use_instance_template ? {} : {
for pair in setproduct(var.iam_roles, keys(local.names)) :
“${pair.0}/${pair.1}” => { role = pair.0, name = pair.1 }
names = (
var.use_instance_template ? { ( = 0 } : {
for i in range(0, var.instance_count) : format(“${}-%04d”, i + 1) => i
service_account_email = (
? (
length(google_service_account.service_account) > 0
? google_service_account.service_account(0).email
: null
: var.service_account
service_account_scopes = (
length(var.service_account_scopes) > 0
? var.service_account_scopes
: (
? (“”
zones_list = length(var.zones) == 0 ? (“${var.region}-b”) : var.zones
zones = {
for name, i in local.names : name => element(local.zones_list, i)

resource “google_compute_disk” “disks” {
for_each = var.use_instance_template ? {} : local.attached_disks_pairs
project = var.project_id
zone = local.zones(
name = each.key
type = local.attached_disks(each.value.disk_name).options.type
size = local.attached_disks(each.value.disk_name).size
image = local.attached_disks(each.value.disk_name).image
labels = merge(var.labels, {
disk_name = local.attached_disks(each.value.disk_name).name
disk_type = local.attached_disks(each.value.disk_name).options.type

# Disk images usually have slashes, which is against label
# restrictions
# image     = local.attached_disks(each.value.disk_name).image

dynamic disk_encryption_key {
for_each = var.encryption != null ? (“”) : ()

content {
  raw_key           = var.encryption.disk_encryption_key_raw
  kms_key_self_link = var.encryption.kms_key_self_link


locals {
snapshot_policy_name = “${var.region}-${var.project_id}-${}-default”

resource “google_compute_disk_resource_policy_attachment” “snapshot_attachments” {
for_each = var.use_instance_template ? {} : local.attached_disks_pairs
project = var.project_id
zone = local.zones(
name = local.snapshot_policy_name
disk = google_compute_disk.disks(each.key).name
depends_on = ( google_compute_resource_policy.snapshot_policy )

resource “google_compute_resource_policy” “snapshot_policy” {
count = var.use_instance_template ? 0 : 1
#for_each = var.use_instance_template ? {} : local.attached_disks_pairs
project = var.project_id
region = var.region
name = local.snapshot_policy_name
snapshot_schedule_policy {
schedule {
daily_schedule {
days_in_cycle = 1
start_time = “09:00”
retention_policy {
max_retention_days = 15
on_source_disk_delete = “KEEP_AUTO_SNAPSHOTS”
snapshot_properties {
storage_locations = (“us”)
guest_flush = false

resource “google_compute_instance” “default” {
for_each = var.use_instance_template ? {} : local.names
project = var.project_id
zone = local.zones(each.key)
name = each.key
hostname = var.hostname
description = “Managed by the compute-vm Terraform module.”
tags = var.tags
machine_type = var.instance_type
min_cpu_platform = var.min_cpu_platform
can_ip_forward = var.can_ip_forward
allow_stopping_for_update = var.options.allow_stopping_for_update
deletion_protection = var.options.deletion_protection
enable_display = var.enable_display
labels = var.labels
metadata = merge(
var.metadata, try(element(var.metadata_list, each.value), {})

lifecycle {
ignore_changes = (

dynamic attached_disk {
for_each = {
for resource_name, pair in local.attached_disks_pairs :
resource_name => local.attached_disks(pair.disk_name) if == each.key
iterator = config
content {
device_name =
mode = config.value.options.mode
source = google_compute_disk.disks(config.key).name

boot_disk {
initialize_params {
type = var.boot_disk.type
image = var.boot_disk.image
size = var.boot_disk.size
disk_encryption_key_raw = var.encryption != null ? var.encryption.disk_encryption_key_raw : null
kms_key_self_link = var.encryption != null ? var.encryption.kms_key_self_link : null

kubernetes – Schedule pod on a node and access pv on another node

I’m running a k3s cluster on RPi4, with heterogenous config (a node has a high capacity but slow hdd, another has a ssd drive, a third only has a sd card).

I have persistent volumes & claims of kind “local-path”, attached to nodes & pods depending on my needs.

I’m facing a situation where I need to schedule a pod on the node with no disk to process data stored in the node with the ssd disk (re-encode some video files to mp4 using ffmpeg, and as this is an expensive process I’d like to do this on an idle node and not slow the node running the ssd).

Is it possible to transparently mount a PV from a different node ? Do I need to use some nfs ? Is there a more evoluted type of volume that can be used in bare-metal RPi4 to do what I want ?

Looking at the docs didn’t help much (there is tons of different persistent volume type, with not many use-case described).


Algorithm to schedule employees into days

I’m trying to build a scheduling app for a friend, but am stuck on how to sort the employees.

I have three holidays each with their own employee_need:

thanks_giving: 2

christmas: 3

new_years_eve: 2

I have employees who have a predetermined number of days they will work. The sum of the employee’s predetermined work days will always add up to the sum of the three holiday’s employee_need. They also have ranked the holidays by preference, which should guide the scheduling process.
The data looks something like this:


days_to_work: 1

preferences: [christmas, new_years_eve, thanks_giving]


days_to_work: 2

preferences: [thanks_giving, christmas, new_years_eve]


Right now my process of sorting is to

  1. Fill each holiday with a list of all employees.

  2. Loop through the employees, starting with those who have the most days off.

  3. Loop through the employee’s preferences and pull them from the one they most desire to have off that also has room for them to be taken off

  4. Continue looping until the days are properly scheduled. If there is room to remove them from that day, I do so, until they are working the number of days they are supposed to.

The algorithm works a decent amount of time, but I really need it to work all the time.

Is anyone familiar with this kind of problem, and can point me toward a better methodology?

windows 10 – How to automate employees schedule

i want to create an excel file with all my employees (from 1 to 24) and all the workdays (from B (01.0.2021) to NB (31.12.2021)).
Every workday i need 5 employees on phone duty so i would like to have some ramdomizer that chooses 5 people for that day. The next day the randomizer chooses another set of 5 people but the 5 from yestersday should be excluded and so on and so forth.
ideally it will generate a sort of calender where people can see if they are on phone duty or not. somewhat like in the picture (the picture shows our current situation, not automated/randomized)
Current Excel

I hope my question is understandable. I’m not very good with excel and dont even know where to start.
any help would be much appreciated.

how do I if know my schedule on crontab is working?


          • sudo java -jar Web-0.0.1-SNAPSHOT.jar

reward schedule – Does adding more miners create more bitcoin?

I was on another forum and the claim was made that the miners using new technology GPUs

GPU mining has been obsolete for Bitcoin mining since around 2014. All mining these days is done using custom chips specifically designed for Bitcoin mining. It is orders of magnitude more power-efficient.

and because of this there will be more bitcoin produced.

Bitcoin has a fixed inflation schedule: 1 block every 10 minutes, and every block currently permits a subsidy 6.25 BTC (a number that halves every 4 years). The difficulty of producing a block goes up (or down) automatically as hashrate is added (or removed). This has a slight delay, so if the hashrate is continuously increasing, blocks will be produced slightly faster than 1 per 10 minutes, as the difficulty will always be slightly behind. This does increase the rate of inflation while it lasts, making the total supply be permanently ahead of schedule. However, due to exponentially decreasing subsidy, the limit is and remains approximately 21 million, regardless of block production speed.

My understanding of block rewards is it’s like a horserace between miners (or group of miners) to solve the blocks and the 1st to solve gets all of the reward.

Who is correct?

Block mining is like a lottery, where every hash tried is one ticket. The lottery is always sold out, so there will be always one winner. There is a new lottery on average every 10 minutes.

It is not like a race; a race would imply that the fastest miner always wins. This is not true, everyone wins approximately proportional to what percentage of the network rate they have.