mySphere Posts


When you create mail journaling you can encrypt all documents using the Administrator ID.  Last week a customer ask to add access for another user (developer). The Admin remove the database encryption but lots off documents remain encrypted . The solution was to remove the document encryption using the following steps:

To remove encryption from documents:

1. First, create a view named “Encrypted”. The view selection formula you use will vary depending on the type of database with which you are working.


  • If working in a mail file, then use the following:
      • SELECT (Encrypt = “1” |  EncryptionFlags = “2”)

    If you are working in a non-mail file and want to remove encryption from documents encrypted with a Private key, then use the following formula:

      • SELECT @IsAvailable(SecretEncryptionKeys)

    If you are working in a non-mail file and documents may be encrypted using either a Public or Private key, then use the following:

      • SELECT @All

    In the above scenario, if you want to remove encryption based on the Form or other field value, then remove the @All. For example: Select Form = “Report”.

    In other scenarios, use an ampersand (&) to add additional criteria to the selection formula. For example: Select Form=”Report” & @IsAvailable(SecretEncryptionKeys)

  • 2. Create an agent with the following code. Set it to run manually from Actions menu on all documents in the database.

Dim s As New notessession
Dim db As notesdatabase
Dim view As notesview
Dim doc As notesdocument
Dim nextdoc As notesdocument

Set db = s.currentdatabase
Set view = db.getview(“Encrypted”)
Set doc = view.getfirstdocument

While Not doc Is Nothing

  • Set nextdoc = view.getnextdocument(doc)
    'The below loop is mandatory to ensure that all $File entries are unecrypted
    Forall i In doc.items

    • If i.isencrypted Then
      • i.isencrypted=False

      End If

    End Forall
    'Must have at least 1 field encrypted in order to call Encrypt method
    Dim temp As New NotesItem(doc,"tempjunk","temp")
    Call doc.encrypt
    Call, False)
    'This portion can now remove the fields relative to encrypting the
    'single token encrypted field.
    Call doc.removeitem("$Seal")
    Call doc.removeitem("$SealData")
    Call doc.removeitem("SecretEncryptionKeys")
    Call doc.removeitem("Encrypt")
    Call doc.removeItem("tempjunk")
    Call, False)
    Set doc = nextdoc



When using the Domino Administrator client, the Rules tab does not display when you select Messaging Settings on the Configuration tab.  You may also notice that after a Configuration document is edited from the Domino Administrator client, the Rules tab no longer appears when the document is viewed from the Notes client.


The issue occurs because the “ActForm” field is present in the configuration document.


To resolve the issue, delete the ActForm field from the Configuration document. There are two approaches to accomplish this:

Approach 1: Create a Formula agent, designed to act on selected documents, to delete the field ActForm:

Steps to create the agent:
1. Start the Domino Designer client and open the Domino Directory database.
2. From the menu, select Create -> Agent.
3. The Agent Properties dialog will appear.  Give the agent a name.  The Runtime settings can be left at the defaults of “Action Menu Selection” and “All Selected Documents”.
4.  Click on the X mark in the upper righthand corner to close and save the Agent Properties settings.  5.  In the main pane, change the “Run” keyword selection from “Simple Action(s)” to “Formula”.
6.  Enter the following code into the main window:
FIELD ActForm :[email protected]
7.  From the menu select File -> Save.
8.  Exit from the Domino Designer client.

Steps to run the agent:
1. Open the Domino Directory from the Notes client.
2. Open the view Configuration -> Servers -> Configurations.
3. Select the affected configuration documents.
4. From the menu select Actions -> <agent name>.

Now when the Configuration document is opened from the Notes client, the Rules tab will display as expected.


Approach 2: Code can be added to the ServerConfig form so that it deletes the field when the Configuration document is opened.

1. Start the Domino Designer client and open the Domino Directory database.
2. In the left pane, select Forms.
3. In the right pane find and open the Form “Server\Configuration Settings”.  Note: The form’s Alias is “ServerConfig”.
4. In the lower left pane the Objects tab will be selected by default, locate and select the Queryopen event listed.
5. Paste the following code into the Programmer’s Pane:

Sub Queryopen(Source As Notesuidocument, Mode As Integer, Isnewdoc As Variant, Continue As Variant)
Dim doc As NotesDocument
Set doc = Source.Document
If Not doc.Isnewnote Then
If doc.HasItem(“ActForm”) Then
Call doc.RemoveItem(“ActForm”)
Call, True)
End If
End If
End Sub

6. From the menu select File->Save.
7. Exit from Domino Designer client.

Now when the Configuration document is opened from the Notes client, the Rules tab will display as expected.




​If you have a problem, and no-one else can help, and if you can find them, maybe you can ask an HCL Ambassador…

See the HCL Ambassadors Class of 2021

Connections Domino Notes WebSphere Portal


When Red Hat, CentOS’s Linux parent company, announced it was “shifting focus from CentOS Linux, the rebuild of Red Hat Enterprise Linux (RHEL), to CentOS Stream, which tracks just ahead of a current RHEL release.” Many CentOS screamed in protest on social media. CentOS co-founder, Gregory Kurtzer, heard them and announced he’d create his own RHEL clone and CentOS replacement: Rocky Linux

Read more information here



Red Hat, CentOS’s Linux parent company, announced it was “shifting focus from CentOS Linux, the rebuild of Red Hat Enterprise Linux (RHEL), to CentOS Stream, which tracks just ahead of a current RHEL release.” In other words, CentOS will no longer be a stable point distribution but a rolling release Linux distribution. CentOS users are ticked off.

Read the full article here



The programming languages used to build the web often find their way into apps, too. That’s largely due to software that allows developers to “reuse” the code they write for the web in products they build to run on operating systems like Linux, Android, Windows, and macOS.

See the full article here



You can now package and deploy AWS Lambda functions as a container image of up to 10 GB.

This makes it easy to build Lambda based applications using familiar container tooling, workflows, and dependencies. Just like functions packaged as ZIP archives, functions deployed as container images will benefit from AWS Lambda’s operational simplicity, automatic scaling with sub-second startup times, high availability, and native integrations with 140 AWS services.

More information here



IBM will begin sunsetting IBM Watson™ Personality Insights on 1 December 2020. For a period of one year from this date, you will still be able to use Personality Insights. However, as of 1 December 2021, the offering will no longer be available.

As an alternative, we encourage you to consider migrating to IBM Watson™ Natural Language Understanding, a service on IBM Cloud that uses deep learning to extract data and insights from text such as keywords, categories, sentiment, emotion, and syntax to provide insights for your business or industry.

More information on this link



With new version of macOS released which is macOS Big Sur 11.0.x, the latest version of our VCP driver for Macintosh OSX (v5.3.5) now becomes incompatible.

The reason for this is because of new driver model change on macOS which you can find more information from links below:

We currently are working to update our driver to fix this issue, although due to testing requirements, we do not expect the new version of driver to be available on our website for now. So please be up to date on our page for VCP drivers below for latest information regarding to the release of the updated VCP driver for macOS Big Sur 11.0.x:

My upgrade to Big Sur will not be possible without this driver. 🙁



A device that has a temperature sensor in an office might report temperature in degrees Fahrenheit or in degrees Celsius. It is not efficient to configure applications to be able to consume data in all these formats. Instead, the data needs to be collected, transformed, and normalized to create a single logical model so that an application can interact with the different devices in the same way.

The data management component of Platform Service includes a device twin feature and an asset twin feature. The device twin feature lets you take advantage of the collection, transformation, and normalization of different formats of device data into a single logical model. The asset twin feature lets you group different devices together to create a Thing, which is a higher value asset-based data structure. You can even group Things together to create new Things. An application can interact with the logical model, regardless of the data format that is used by the individual devices or Things.

For example, a group of devices that report temperature, humidity, and ambient light can be aggregated into a “Room” Thing to represent the comfort level in a specific office. A number of “Room” Things can be aggregated into a “Floor” Thing to represent all offices on a specific level, and a number of “Floor” Things can be aggregated into a “Building” Thing. By using a Thing abstraction, your application is decoupled from the specifics of how the devices are connected, the format in which the devices publish event data, and how the data is combined.

The explanation above i got from the knowledge center, just to  introduce about a problem i  got last week.

The problem:  When we create device types  we must define a Hardware and a Logical interface. When we create  the interfaces we are modeling the schema of a table were data will be saved. You can define “the fields and his type”. As usual i create my interfaces, devices and put the sensors to send data, but some sensors are not sending one of the variables. For example the data expected was { voltage : value , kwh: value) but the sensor send {voltage: value}.  The data is not saved on the database.

To solve the problem i create a ticket and the answer was to test if a key exists if not save it as a null value.

On the logical interface, for each key i need to put the following (example for kwh) : $exists($event.kwh) ? $event.kwh(null)





I setup Node-red to use HTTPS and everything works fine. All HTTPS requests go to default port of Node-red.  Today i need to setup a communication with a client that not support HTTPS, just HTTP.

The solution i found was to setup Nginx as reverse proxy.  Nginx receives the HTTP requests and then forward them to Node-red.

On my Nginx server i just add the following lines

server {
listen 80;
location /reddata {
proxy_pass https://<ipaddress>:port/reddata;



IoT Linux