Friday, January 31, 2025

How to turn off or reset MFA from Microsoft admin console

To turn off or reset Multi-Factor Authentication (MFA) from the Microsoft admin console, follow these steps:

Turning Off MFA:
  1. Sign in to the Azure portal as a global administrator.

  2. Navigate to Azure Active Directory > Properties.

  3. Select Manage security defaults.

  4. Set Security defaults to Disabled and click Save.

Resetting MFA for a User:
  1. Sign in to the Microsoft 365 admin center with your admin credentials.

  2. Go to Users > Active users.

  3. Select the user who needs the MFA reset.

  4. Click on Reset multi-factor authentication under More settings2.

  5. Follow the prompts to reset the user's MFA settings.

Disabling MFA for a Specific User:
  1. Sign in to the Microsoft 365 admin center.

  2. Go to Users > Active users.

  3. Select the user you want to disable MFA for.

  4. Click on Manage multifactor authentication.

  5. Select the user and click Disable multi-factor authentication3.

These steps should help you manage MFA settings effectively!!

Saturday, January 25, 2025

What are advantages of Pinecone? Why Pinecone?

Pinecone is a powerful vector database designed to accelerate AI applications. Here's why it's worth considering:

  1. Vector Search: Pinecone represents data as vectors, allowing it to quickly search for similar data points in a database. This makes it ideal for various use cases, including semantic search, similarity search for images and audio, recommendation systems, record matching, and anomaly detection1.

  2. Managed and Cloud-Native: Pinecone is a managed service, meaning you don't have to worry about infrastructure hassles. It serves fresh, relevant query results with low latency, even at the scale of billions of vectors2.

  3. Serverless: Pinecone is serverless, which simplifies scaling and management. You can create an account, set up an index, and upload vector embeddings in just 30 seconds1.

Whether you're building recommendation engines, search systems, or anomaly detectors, Pinecone can help power your AI applications efficiently.

Thank You!!

Saturday, January 18, 2025

Understanding Overfitting and Underfitting in Machine Learning

In the realm of machine learning, overfitting and underfitting are common challenges that impede the performance of models. These issues are central to the capacity of a model to generalize well, ultimately affecting its usefulness in providing accurate and reliable predictions.

 

What is Overfitting and Underfitting?

Before delving deep into the implications of overfitting and underfitting, it's crucial to comprehend several fundamental concepts that underpin these phenomena. The terms "signal" and "noise" are pivotal in understanding the behaviour of machine learning models. Signal refers to the true underlying pattern of data that facilitates learning, while noise encompasses irrelevant and extraneous data that diminishes performance.

Similarly, bias and variance play crucial roles in model evaluation. Bias signifies the prediction error arising from oversimplifying the learning algorithm, whereas variance occurs when the model performs well with the training data but struggles with the test data.

 

Overfitting: An In-Depth Analysis

Overfitting transpires when a machine learning model endeavours to encapsulate all data points within the dataset, even to the extent of accommodating more information than necessary. This results in the model capturing noise and inaccuracies from the data, thereby undermining its efficiency and accuracy. Overfitted models often exhibit low bias and high variance, signifying their susceptibility to deviate markedly from the expected outcome.

A classic example of overfitting can be comprehended through a linear regression output, wherein the model rigorously attempts to envelop all data points, thereby resulting in suboptimal performance and prediction errors.

 

Mitigating Overfitting: Techniques and Strategies

To obviate the menace of overfitting, a slew of techniques can be employed, including cross-validation, augmenting the training dataset, feature selection, early stopping, regularization, and ensembling. These strategies are aimed at instilling a sense of balance and generalization within the model, thereby rectifying the aberrations stemming from overfitting.

Understanding Underfitting and Counteracting It

Conversely, underfitting occurs when a machine learning model fails to grasp the underlying trend inherent within the data. This phenomenon can unfold when the model is prematurely halted during the training phase, impeding its ability to discern patterns and relationships from the data. Models afflicted by underfitting exhibit high bias and low variance, ultimately leading to unreliable and inaccurate predictions.

An illustration of underfitting can be elucidated through a linear regression model output, where the model's inability to encapsulate the data points reflects its inadequacy in learning from the dataset.

 

Strategies to Combat Underfitting

To avert underfitting, measures such as prolonging the training duration and augmenting the number of features can be instrumental. These actions are designed to empower the model to learn comprehensively from the training data, thereby fostering an enhanced capacity to discern and encapsulate the dominant trend within the dataset.

 

Striving for Goodness of Fit

The ultimate ambition of machine learning models is to achieve a state of goodness of fit, where the model strikes a harmonious equilibrium between underfitting and overfitting. This state implies that the model is capable of making predictions with minimal errors, thus epitomizing the essence of generalization.

There are several methods to discern and attain the stage of goodness of fit, including resampling techniques to estimate model accuracy and the deployment of validation datasets.

 

Final Thoughts

The perils of overfitting and underfitting are ubiquitous in the realm of machine learning, underscoring the need for robust strategies and techniques to mitigate their deleterious impact. By leveraging a judicious combination of model evaluation, feature engineering, and regularization, machine learning practitioners can navigate these challenges and foster models that exude resilience, precision, and reliability.

Sunday, December 01, 2024

How to find specific table or view is used in SQL Server database

Here is how we can find a table or view in SQL database. Below query will help you find the table or view and which object has used.

select schema_name(o.schema_id) + '.' + o.name as [table],
       'is used by' as ref,
       schema_name(ref_o.schema_id) + '.' + ref_o.name as [object],
       ref_o.type_desc as object_type
from sys.objects o
join sys.sql_expression_dependencies dep
     on o.object_id = dep.referenced_id
join sys.objects ref_o
     on dep.referencing_id = ref_o.object_id
where o.type in ('V', 'U')
      --and schema_name(o.schema_id) = 'dbo'  -- put schema name here
      and 
	  o.name = 'AI_Asset_Info'   -- put table/view name here
order by [object]
  

Hope this helps!

Tuesday, November 26, 2024

How to Copy Git Repository Without History

There are several methods to do this using git clone, git push or using git archive. But I personally prefer the one using git clone.

Objective is to copy repo 1 which is source repo to a new Repo which is NewRemote repo with out commit history.

Precautions before you proceed with this:

  1. Ensure you have write access to the repository.
  2. Backup any important local changes before proceeding.
  3. This will permanently remove the old commit history.
  4. Collaborators will need to re-clone the repository.

Here are step by step git examples for this specific repo

# 1. Clone the source repository
git clone https://github.com/inagasai/SourceRepo.App.git

# 2. Enter the cloned repository directory
cd vGlence.App

# 3. Verify current branches
git branch -a

# 4. Checkout master branch
git checkout master

# 5. Create a new branch without history
git checkout --orphan clean-main

# 6. Add all files to the new branch
git add .

# 7. Commit the files with a new initial commit
git commit -m "Initial commit - reset repository history"

# 8. Delete the old main branch (if it exists)
git branch -D main 2>/dev/null

# 9. Rename current branch to main
git branch -m main

# 10. Remove the original remote
git remote remove origin

# 11. Add the original repository as a new remote
git remote add origin https://github.com/inagasai/NewRemote.App.git

# 12. Force push to overwrite the remote repository
git push -f origin main
  

Detailed Breakdown of the outcome:

  1. This process creates a new branch with no commit history.
  2. It adds all existing files to a new initial commit.
  3. Force pushes to overwrite the remote repository.
  4. Removes all previous commit history.

Hope this helps.